Abstract

With the recent rise in artificial intelligence, several areas of computer science have begun to explore the use of machine learning algorithms. Among them, recent video encoding techniques exploit in large-scale machine learning algorithms, either to aid in compression efficiency or to reduce computational cost. However, this is not a simple task, especially if the objective is to reduce computational cost, since it is necessary that the machine learning algorithm used does not interfere negatively in the total coding process time and in compression efficiency. As video transrating is an essential task in streaming service providers that need to transmit and deliver different versions of the same content for a multitude of users operating under different network conditions, it is critical to find an efficient strategy that yields computational cost reduction with small loss in compression efficiency. This work discusses the use of different machine learning algorithms in the video transcoding concept, comparing them in terms of accuracy and overhead processing time. The experimental results and discussions provide insights that can be useful in the definition of future fast transcoding schemes in real-world transcoding architectures.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.