Abstract
A Robust Adaptive Video Encoder (RAVE) based on human visual model is proposed. The encoder combines the best features of Fine Granularity Scalable (FGS) coding, frame-dropping coding, video redundancy coding, and human visual model. According to packet loss and available bandwidth of the network, the encoder adjust the output bit rate by jointly adapting quantization step-size instructed by human visual model, rate shaping, and periodically inserting key frame. The proposed encoder is implemented based on MPEG-4 encoder and is compared with the case of a conventional FGS algorithm. It is shown that RAVE is a very efficient robust video encoder that provides improved visual quality for the receiver and consumes equal or less network resource. Results are confirmed by subjective tests and simulation tests.
References
Weiping Li, MPEG-4 Video Standard, IEEE Trans. on Circuits and Systems for Video Technology, 11(2001)3, 301–317.
Eric C. Reed, Frederic Dufaux, Constrained bit-rate control for very low bit-rate streaming-video application, IEEE Trans. on Circuits and Systems for Video Technology, 11(2001)7, 882–889.
N. Jayant, J. Johnston, R. Safranek, Signal compression based on models of human perception, Proc. IEEE, 81(1993)10, 1385–1422.
N. Jayant, Signal compression: Technology targets and research directions, IEEE J. on SAC, 10(1992)7, 796–818.
C. H. Chou, C. W. Chen, A perceptually optimized 3-D subband codec for video communication over wireless channels, IEEE Trans. on Circuits and Systems for Video Technology, 6(1996)4, 143–156.
N. Jayant, Signal compression: Technology targets and research directions, IEEE J. on SAC, 10(1992)7, 796–818.
Author information
Authors and Affiliations
Additional information
Supported by Innovation Fund of China(00C26224210641)
About this article
Cite this article
Yin, H., Zhang, J., Zhu, Y. et al. A robust adaptive video encoder based on human visual model. J. of Electron.(China) 20, 142–149 (2003). https://doi.org/10.1007/s11767-003-0011-0
Received:
Revised:
Issue Date:
DOI: https://doi.org/10.1007/s11767-003-0011-0