Professional Documents
Culture Documents
IJRAR1CSP053
IJRAR1CSP053
Keywords Abstract
Video A fresh comparison analysis on video compression technologies was offered in this study. Video
Compression, streaming programmes are becoming increasingly popular as internet technology and computers
Algorithm, advance at a rapid pace. As a result, uncompressed raw video demands a lot of disc space and
Deep Learning , network bandwidth to store and deliver today. We describe a novel technique to video compression
Motion Estimation, surveillance that improves on the shortcomings of previous approaches by replacing each standard
BitRate Estimation. component with a neural network counterpart. . We describe a novel technique to video
compression surveillance that improves on the shortcomings of previous approaches by replacing
each standard component with a neural network counterpart. It delivers a higher-quality video
stream at a consistent bit rate (compared to previous standards). As a result, you must select the
appropriate video compression technology to fulfil the requirements of your video application.. Our
work is founded on a set of principles common method for reducing the bit rate while minimising
distortions in decoded frames by using In video frames, there is spatial and temporal redundancy.
We use a neural network to develop a video compression in the traditional sense strategy and
encode redundant data with fewer bits. Experiments have revealed that our solution is successful
and surpass traditional MPEG encoding while retaining visual quality at similar bit rates. Although
our approach is geared at surveillance, it can simply be applied to other types of video.
IJRAR1CSP053 International Journal of Research and Analytical Reviews (IJRAR) www.ijrar.org 349
© 2022 IJRAR August 2022, Volume 9, Issue 3 www.ijrar.org (E-ISSN 2348-1269, P- ISSN 2349-5138)
METHODOLOGY USED own receptive field, which is the area where the neuron receives
information. Linked to neighbouring neurons' receptive areas
A. TRANSFORMATION FUNCTIONS to encompass the whole visual field. Only the biological
perception system's neurons are involved. Responds to external
The frame data is converted from the spatial domain using stimuli Each neuron in a CNN analyses data solely in its
the transformations function. This thesis focuses on the most receptive area, which is a small portion of the brain. The
recent and widely discussed innovations in the last five years. receptive field is a small part of the visual field. A smaller
As a result, H.264 and HEVC have been chosen for further
amount of complicated pattern (lines, curves, and so on) appear
investigation. The transform basis functions are obtained
from the DCT in video coding techniques in general and initially in the layers, then in the more intricate pattern (faces,
H.264, HEVC in particular. The frequency domain is where objects, etc.). A CNN can be utilised to give a computer sight.
redundancy can be found, and the spatial domain is where The convolution layer is the most important part of the CNN. It
image component picture elements are transformed using the is in charge of the majority of the system's processing power.
DCT approach. In video compression, the frame is split up
into blocks that range in size from 4*4 to 64*64. a picture- This layer does a dot product of two-dimensional matrices, the
compression technique The image is divided into 8*8 blocks restricted matrix being one of them. area of the receptive field
by JPEG, which then performs a two-dimensional discrete
and the other a set of learnable parameters called a kernel.
cosine transform (DCT) on each of the blocks.
Despite its modest size, the kernel has more depth than a picture.
B.QUANTIZATION FUNCTION If the image has a small size, the kernel height and width will be
small. Although there are three (RGB) channels, the depth will
The inevitability of representing a value as a number with a span all three. A neural network could also include up to 20 or
definite number of decimal places is quantization. The 30 layers. The convolutional layer is a special element of layer.
quantization scale code is divided element-wise by the Several layers of a convolutional neural network is what really
quantization matrix in video coding, and each resultant give it its power.
element is rounded. The step size for associating the
transformed coefficients with a finite number of steps is
determined by the quantization parameter.When it comes to
video coding, the value is inversely proportional to the PSNR RESULT AND DISCUSSION
value and directly proportional to the CR. In both cases, the
DC value is the value with zero frequency. Inverse In our dataset, the average PSNR and MS-SSIM values were
quantization formulas are applied to a converted signal, and 74 dB and 0.98, respectivelyWhen the frames are similar to
equations (3) and (4) are the quantization equations. the frames in the dataset, our technique outperforms standard
Equations (5) and (6), on the other hand, are the quantization MPEG in terms of MS-SSIM and PSNR and is comparable to
formulas for intra and inter-coding with AC values the non- standard H.264 in terms of MS-SSIM.
zero frequencies. The inverse quantization formula used on
the Standard is equation (7). Despite being trained on MS-SSIM, our model performs
better in terms of PSNR than the typical MPEG model. But
more crucially, We see that at the lower bitrate, the
reconstructed frames' visual quality has increased.
Due to its smaller size and fewer parameters, our model runs
quicker and uses less memory. Due to its exact correspondence
to the conventional video compression technique, our proposed
framework can serve as a reference model for deep learning-
based video compression.
Fig. Comparision betn our model
C.ENTROPY CODING
IJRAR1CSP053 International Journal of Research and Analytical Reviews (IJRAR) www.ijrar.org 350
© 2022 IJRAR August 2022, Volume 9, Issue 3 www.ijrar.org (E-ISSN 2348-1269, P- ISSN 2349-5138)
1811.06981.
[8] Farnebäck, G. (2003). Two-frame motion estimation based
on polynomial expansion, In Scandinavian conference on
image analysis. Springer.
CONCLUSION
ACKNOWLEDGEMENT
REFERENCES