lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <4705F658.1050106@joow.be>
Date:	Fri, 05 Oct 2007 10:31:20 +0200
From:	Pieter Palmers <pieterp@...w.be>
To:	Kristian Høgsberg <krh@...planet.net>
CC:	Stefan Richter <stefanr@...6.in-berlin.de>,
	Kristian Høgsberg 
	<krh@...hat.com>, linux1394-devel@...ts.sourceforge.net,
	linux-kernel@...r.kernel.org
Subject: Re: [PATCH] firewire: adopt read cycle timer ABI from raw1394

Kristian Høgsberg wrote:
> On 10/1/07, Pieter Palmers <pieterp@...w.be> wrote:
>> Stefan Richter wrote:
>>>> This duplicates the read cycle timer feature of raw1394 (added in Linux
>>>> 2.6.21) in firewire-core's userspace ABI.
>>> Kristian and Pieter, does this simple duplication of the ioctl make
>>> sense on its own?  AFAIU rawiso's iso packet buffers look different from
>>> fw-cdevs's. It seems to me as if rawiso always put the cycle into a user
>>> buffer for each iso packet received...
>>>
>>> raw1394.h::struct raw1394_iso_packet_info {
>>>       __u32 offset;
>>>       __u16 len;
>>>       __u16 cycle;   /* recv only */
>>>       __u8  channel; /* recv only */
>>>       __u8  tag;
>>>       __u8  sy;
>>> };
>>>
>>> raw1394.c::raw1394_iso_recv_packets()
>>>
>>>       /* copy the packet_infos out */
>>>       for (i = 0; i < upackets.n_packets; i++) {
>>>               if (__copy_to_user(&upackets.infos[i],
>>>                                  &fi->iso_handle->infos[packet],
>>>                                  sizeof(struct raw1394_iso_packet_info)))
>>>                       return -EFAULT;
>>>
>>>               packet = (packet + 1) % fi->iso_handle->buf_packets;
>>>       }
>>>
>>> ...while the Juju ABI returns the cycle only for those packets whose
>>> fw_cdev_iso_packet.control had the FW_CDEV_ISO_INTERRUPT flag set.
>>> The cycle is then written out in the fw_cdev_event_iso_interrupt event
>>> which happens when this particular packet was received.  Right?
>>>
>>> Pieter, do applications like yours need the cycle counter only for a few
>>> predetermined packets or for each and every packet?
>> We need it for every packet for two reasons:
>> 1) it's the only way to determine how many packets were dropped when
>> packet drops are flagged in the callback
> 
> Your application should know what the timestamp should be for each iso
> receive callback and if you see a larger value than expected you can
> use that to calculate how many cycles were lost.
I'm not convinced here. It's rather easy to come up with a scenario 
where more than 16 packets are lost, hence the timestamp wraps around 
and you have the impression that no packets are lost. Losing 16 packets 
is a serious issue in streaming audio, so we definitely have to be able 
to detect this.

The point is that currently we use the cycle value for every callback 
since that's how the libraw API is defined. So if our current code is 
supposed to work with the new kernel modules through the adapted libraw 
it obviously has to provide the cycle parameter for every callback. 
Therefore I suppose the discussion is about a new API that is being 
designed.

> 
>> 2) we convert the 16-bit SYT timestamp of a packet to a full 32-bit
>> cycle counter value. This because the range of the 16-bit SYT is too
>> small (only 16 packets) for systems that have large buffering.
> 
> If you get the timestamp for the first packet in a receive batch, you
> can still do this, right?

If we can calculate the cycle a packet is received on one way or 
another, we are ok. So whether this is done by receiving a cycle value 
for each packet, or having the cycle for the first packet in a batch and 
then using offset calculations it is ok. The only precondition for this 
is that a batch should be continuous, i.e. there should not be any 
missing packets in a batch. Only between batches.

We will probably only be able to comment on this thoroughly once we 
start implementing the juju based streaming system. I don't see this 
happening in the next 6 months though.

Greets,

Pieter

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ