Hi,
I'am trying to understand why everywhere the issue with audio-source-to-SPDIF conversion is so much sensitive discussed when it comes to quality. Its always mentioned that jitter sensitivity lies in the picosecond area! Comparing this latency with common network latency in the typical homeuser network, which lies around 0.2-3 milliseconds when wires are used or 0.5-5 milliseconds if a good wireless network is used, I'am wondering how much truth lies in that, as 1 second is 1.000 milliseconds and 1 millisecond is 1.000.000.000 picoseconds (what the hell!).
When thinking about digital streaming I'am thinking about sending audio (PCM or whatever) as RTP (real-time transport protocol) over the network to the RTP stream-receiver or playback device (eg. Apple Airport Express). Other possibilities are mounting filesystems over the network like NFS or SMB and reading the audio data with filesystem operations (opening files over network and reading them). In my understand it should be unimportant which kind of mechanism is used as all the audio data should be cached in the playback device for a small amount of time - so in case of RTP the packets are reordered first and then cached for final playback - and after the caching interval (time or data amount) the sampling of the data to SPDIF-out should begin.
Using that basic principle the way network streaming works why should network latency (when no buffer/cache underrun happens) influence the quality of the SPDIF-out sampling at all? Shouldn't be the only mattering parameter the SPDIF-to-SPDIF jitter (e.g. networkplayer-to-AVRorDAC)? Shouldn't that jitter be deduced from the network side to the SPDIF side?
I'am trying to understand why everywhere the issue with audio-source-to-SPDIF conversion is so much sensitive discussed when it comes to quality. Its always mentioned that jitter sensitivity lies in the picosecond area! Comparing this latency with common network latency in the typical homeuser network, which lies around 0.2-3 milliseconds when wires are used or 0.5-5 milliseconds if a good wireless network is used, I'am wondering how much truth lies in that, as 1 second is 1.000 milliseconds and 1 millisecond is 1.000.000.000 picoseconds (what the hell!).
When thinking about digital streaming I'am thinking about sending audio (PCM or whatever) as RTP (real-time transport protocol) over the network to the RTP stream-receiver or playback device (eg. Apple Airport Express). Other possibilities are mounting filesystems over the network like NFS or SMB and reading the audio data with filesystem operations (opening files over network and reading them). In my understand it should be unimportant which kind of mechanism is used as all the audio data should be cached in the playback device for a small amount of time - so in case of RTP the packets are reordered first and then cached for final playback - and after the caching interval (time or data amount) the sampling of the data to SPDIF-out should begin.
Using that basic principle the way network streaming works why should network latency (when no buffer/cache underrun happens) influence the quality of the SPDIF-out sampling at all? Shouldn't be the only mattering parameter the SPDIF-to-SPDIF jitter (e.g. networkplayer-to-AVRorDAC)? Shouldn't that jitter be deduced from the network side to the SPDIF side?