[NUT-devel] Another suggestion for broadcast [PATCH]
Måns Rullgård
mans at mansr.com
Mon Feb 11 13:14:20 CET 2008
Michael Niedermayer wrote:
> On Thu, Feb 07, 2008 at 01:58:32PM -0000, Måns Rullgård wrote:
>>
>> Michael Niedermayer wrote:
> [...]
>> > That way the buffer_fullness stored in the syncpoint will always match
>> > exactly the amount the decoder has in its buffer when reading the
>> > syncpoint. If it has more in its buffer it just would change its clock
>> > to one running at 100.1% and when it has less in its buffer it would
>> > choose a 99.9% clock as reference. (Or any approximetaly equivalent
>> > process)
>>
>> That the buffer fullness is off by N bits doesn't tell you how much too
>> fast or too slow your clock is, only the sign of the error.
>
> yes
>
>
>> Knowing
>> also the magnitude of the error allows much more rapid convergence.
>
> I am not so sure about this. I mean i dont dispute that more information
> should improve it, but i think its good enough with the too much/too little.
>
> A simple example, lets assume we have a decoder with a clock that drifts
> by up to D between syncpoints.
> That is, in the most ideal case we would have to accept that we are
> D off when we reach a syncpoint, assuming we synced to the previous
> perfectly.
>
> Now lets assume that we are within -2D .. +2D at syncpoint x, and we apply
> a +D correction if we are <0 and -D if we are >0. This correction could be
> applied slowly until the next syncpoint. What matters is that after the
> correction we are within -D .. +D and with the drift thats again -2D .. +2D
> at syncpoint x+1.
> Thus above is a proof by induction that just knowing the sign and the worst
> case clock drift is sufficient to be within a factor of 2 of the best
> achiveable clock sync. (comparing worst cases, not average)
This is not how clock sync is usually done. A typical implementation
involves a PLL-type construct to make the local clock accurately track
the sender clock. Once locked, there is very little drift. To correctly
compensate for what little drift inevitably remains, the size of the
error must be known.
The time difference can of course be computed from the difference in buffer
fullness and the received bitrate, it merely takes a little more work on
the receiver side.
>> Providing the timestamp in the stream makes this trivial and independent
>> of the buffering mechanism actually used. Only specifying expected
>> buffer fullness (according to a reference model) requires that the
>> receiver at the very least simulate the reference model,
>
> I think most receivers will use something quite similar to the reference
> model thus making this unneeded. Though yes a receiver using a different
> buffer model might need to simulate the reference one.
I think it very unlikely that any real implementation will use whatever
precise buffer model we choose. Just about any implementation is
likely to immediately extract the elementary streams of interest, and
discard everything else, such as container headers and unwanted elementary
streams.
> But i have difficulty imageing a sufficiently different buffer model.
> I mean a receiver with split buffers could just be taking the sum of
> their buffer_fullness.
> A reciver which removes packets later or not instantaneously would just
> traverse the last few packets to find out how much the refernce buffer
> would contain.
I'm not saying it would very difficult to simulate the reference buffer,
but something is always more than nothing.
> Anyway if you say a timestamp would be better, i suspect rich would be harder
> to convince than me.
Of that I am certain.
--
Måns Rullgård
mans at mansr.com
More information about the NUT-devel
mailing list