Professional Documents
Culture Documents
Project - Video Compression - Scribd
Project - Video Compression - Scribd
Project - Video Compression - Scribd
Most data generated in the world today is videos. With the increasing of video content,
nowadays, 70% of Internet traffic is used for video-based applications, including live
streaming like (YouTube and Twitch), low-latency real time online communication, and
video-on-demand platform (Zoom and Webex). The main task given to compression
techniques is to minimize the number of bits required for code on data or information
provided, further minimizing the memory required to store the given data. Graceful
degradation is a quality-of-service term used to explain that as bandwidth drops or
transmission error occurs, the user experience becomes degraded and tries to be
meaningful. Traditional data compression algorithms use handcrafted encoder–decoder
pairs called “codecs”. Video compression standards such as H.264/AVC, H.265/HEVC,
and the incoming H.266/VVC get involved in those artifacts by assigning large
quantization parameter and coding unit size.
Meanwhile, video resolution and fidelity have also made huge steps forward (e.g., 4k, 8k,
gigapixel, high dynamic range, bit-depth) that make the dire situation worse. Therefore,
many efforts have been invested in improving the video compression algorithms, whose
task is to reduce the video size while keeping an acceptable visual quality reflected in the
human visual system. Moreover, compression is necessary for much real-time and
complex applications such as space, live time-series data, and medical imaging, which
require exact recoveries of original images. Many human efforts are spent analysing the
details of these new data formats and providing efficient compression methods.
LITERATURE REVIEW
• A survey on video compression fast block matching algorithms
Reducing the amount of data required to represent digital video while maintaining
video quality is known as video compression. It consists of multimedia
transmission, video phone, teleconferencing, high definition television, CD-ROM
storage, etc. Compression is a technique used to remove redundant information in
a video sequence. According to (Richardson, 2010), digital video consists of four
types of redundancy, namely color, space, temporal and statistical. Redundant
segregation is processed separately based on feature differences.
As is known, video coding consists of two systems, namely video encoder and
video decoder. The video encoder consists of three main functional units, which
are color subsamples, temporal models (encoders between frames) or spatial
models (encoders within frames) and entropy encoders. The goal of the encoder
is to condense the large amount of information required to display a video frame
to achieve a high compression ratio as shown:
Coder classification can be divided into two approaches namely lossy and lossless.
Lossless is defined as bit preservation or a reversible method used for statistical
redundancy compression. The ratio of this compression method is low which is
about 3:1 or 4:1 in the best case, but the reconstructed data is the same as the
original data. While the lossy technique has a high compression ratio of 50:1 to
200:1 or more than usual. However, the reconstructed data is not the same as the
original data and causes loss of information (H.Mukhtar, A.Al-Dweik, M .Al-Mualla,
2016).
However, the use of block matching algorithms can provide an overview of
solutions based on competitiveness, computational complexity, processing time
and distortion levels. The block matching algorithm can be observed and
emphasized based on a summary of video compression techniques as shown in
table 1..
.
Type of
Compression
Choudhary, S.M.; Patel, U.S.; Parmar, 2015 states that the compression process
is complex and needs to be completed with two Steps, namely decollation and
entropy. Decorrelation removes redundancy between pixels with decorrelation
techniques such as long-term coding, prediction techniques, transformation
techniques or SCAN language-based methodologies. Entropy coding eliminates
coding redundancy. Entropy means how many average bits are required to
represent a symbol. In coding, for frequently used symbols, fewer bits (values less
than entropy) are assigned, and more bits (values greater than entropy) are
assigned to infrequently used symbols. This led to the formation of VLC (Variable
Length Code).
Therefore, identifying intelligent models that can control data in various real-time
applications is very important. For example, in the case of a video call, it may
require seeing a person or people then other objects in the frame. Another example
involves a tennis match. In a tennis match, it is more important to maintain the
quality of the players and the court than the distinguishing features of the audience.
Based on the example, it is clearly proven that the use of machine learning
approaches can meet expectations. Therefore, there are so many algorithms for
machine learning to perform functions such as regression, classification, grouping,
decision trees, extrapolation and more. Machine learning trains algorithms to
extract data information to perform data-dependent tasks. While designing such
algorithms, various machine learning approaches such as Supervised Learning,
Unsupervised Learning, Reinforcement Learning, etc., can be used (Van Eck, N.J.;
Waltman, 2010).
• Joint video compression and encryption using parallel compressive sensing
and improved chaotic maps.
Data compression is possible because images are extremely data intensive and
contain a large amount of redundancy which can be removed by accomplishing
some kind of transform, with a reversible linear phase to de-correlate the image
data pixels (Khalifa and Dlay, 1998). To understand the video formats, the
characteristics of the video and how these characteristics are used in defining the
format need to be understood. Video is a sequence of images which are displayed
in order. Each of these images is called a frame. Since, we cannot notice small
changes in the frames like a slight difference of color, video compression standards
do not encode all the details in the video; some of the details are actually lost
(Abomhara et al., 2010). This is called lossy compression. It is possible to get very
high compression ratios when lossy compression is used. Whereas there are some
compressions techniques are reversible or non-destructive compression (Haseeb
and Khalifa, 2006). It is guaranteed that the decompression image is identical to
the original image. This is an important requirement for some applications where’
high quality is demanded. This called lossless compression (Khalifa and Dlay,
1998, 1999). Typically, 30 frames are displayed on the screen every second. There
will be lots of information repeated in the consecutive frames. If a tree is displayed
for one second, then 30 frames are used for that tree. This information can be used
in the compression and the frames can be defined based upon previous frames.
Frames can be compressed using only the information in that frame (intraframe)
or using information in other frames as well (intraframe).
The first video compression standard to gain widespread acceptance was the
H.261 standard. The H.261 and 263 standards are suitable for carrying video over
ISDN. They are used for video delivery over low bandwidths (Marcel et al., 1997).
The MPEG standards provide a range of compression formats that are suitable for
applications that require higher bit rates. The MPEG-1 provides compression for
standard VHS quality video compression. The MPEG-2 meets the requirements of
applications with bit rates up to 100 Mbps and can easily cater for digital television
broadcasting applications. Video compression is gaining popularity since storage
and network bandwidth requirements can be reduced with compression. Many
algorithms for video compression which are designed with a different target in mind
have been proposed.
CONCLUSION
[3] J. Cai , D. Pan , On fast and accurate block-based motion estimation algorithms
using particle swarm optimization, Int. J. Inf. Sci. 197 (2012) 53–64 .
[4] D.M. Thomas , S. Varier , A novel based approach for finding motion estimation
in video compression, Int. J. Adv. Res. Comput. Commun. Eng. 1 (8) (2012)
514–520 .
[6] Richardson, John Wiley & Sons Inc. (2010): The H.264 Advanced Video
Compression Standard, second ed.
[8] van Eck, N.J. and Waltman, L. (2009). Software survey: VOS viewer, a computer
program for bibliometric mapping. Scient metrics, 84(2), pp.523–538.
doi:10.1007/s11192-009-0146-3. I.E.G. Richardson , The H.264 Advanced Video
Compression Standard, second ed., John Wiley & Sons Inc., UK, 2010
[9] Choudhary, S.M.; Patel, A.S.; Parmar, (2015) .S.J. Study of LZ77 and LZ78
Data Compression Techniques. Int. J. Eng. Sci. Innov. Technology.
[13] X. Zhang, S.-H. Seo, C. Wang, A lightweight encryption method for privacy
protection in surveillance videos, IEEE Access 6 (2018) 18074–18087
[14] J. Sethi, J. Bhaumik, A.S. Chowdhury, Fast and secure video encryption using
divide-and-conquer and logistic tent infinite collapse chaotic map, in: International
Confer-ence on Computer Vision and Image Processing, 2022, pp.151–163.
[17] Khalifa, O.O. and S.S. Dlay, 1998. Wavelets image data compression.
Proceedings of the IEEE International Symposium on Industrial Electronics, July
07, South Africa, pp: 407-410.
[18] Jeremiah, G., 2004. Comparing media codecs for video content. Proceedings of
the Embedded Systems Conference, (ESC`04), San Francisco, pp: 1-18.
[19] Khalifa, O.O., 2003. Image data compression in wavelet transform domain using
modified LBG algorithm. ACM Int. Conf. Proc. Ser., 49: 88-93
[20] Abomhara, M., O. Zakaria, O.O. Khalifa, A.A. Zaidan and B.B. Zaidan,
2010. Enhancing selective encryption for H.264/AVC using advance encryption
standard. Int. J. Comput. Electr. Eng., 2: 1793-8201.
[21] Haseeb, S. and O.O. Khalifa, 2006. Comparative performance analysis of image
compression by JPEG 2000: A case study on medical images. Inform. Technol.
J., 5: 35-39.
[21] Haseeb, S. and O.O. Khalifa, 2006. Comparative performance analysis of image
compression by JPEG 2000: A case study on medical images. Inform. Technol.
J., 5: 35-39.
[22] Marcel, A., H.S. Cornelis and J.B. Ferderik, 1997. Low-bitrate video coding based
upon geometric transformations. Proceeding of the proRISC Work on Circuits,
System and Signal Processing, 20 May 1998, South Africa, pp: 561-568.