Explain why jitter is undesirable for real-time communications such as voice and video communications?
Is this correct? The question says explain, I don't think I've explained as much as stated Jitter causes the received signal to breakup due to the variability of delay. this can lead to voice and video breaking up and becoming unintelligible, which would in turn defeat the purpose of the communication.
Jitter can refer to a number of things. There are a multitude of different noises (jitter) that can occur. Jitter due to thermal noise caused by the hardware, other signals intercepting, inability of the receiver to decode at a certain speed, or even mismatch between the baud rate of the transmitter and the receiver. There are multitudes more. These are the basic common ones. Anywho, this is bad for any type of communication because essentially, jitter is error in the data bits. In real time communication, this can be bad cause your image/voice might be distorted and/or become unintelligible. But I'd assume that for real time communication, the code doesn't have enough time to run a full error correction algorithm on it, and has to accept a higher error threshold. In which case, like you stated it could cause the signal to break up or become fully corrupt.
thanks So I'm not really missing anything in the answer...the answers just rather simplistic, right?
Join our real-time social learning platform and learn together with your friends!