FMUSER Wirless Transmit Video And Audio More Easier !
es.fmuser.org
it.fmuser.org
fr.fmuser.org
de.fmuser.org
af.fmuser.org ->Afrikaans
sq.fmuser.org ->Albanian
ar.fmuser.org ->Arabic
hy.fmuser.org ->Armenian
az.fmuser.org ->Azerbaijani
eu.fmuser.org ->Basque
be.fmuser.org ->Belarusian
bg.fmuser.org ->Bulgarian
ca.fmuser.org ->Catalan
zh-CN.fmuser.org ->Chinese (Simplified)
zh-TW.fmuser.org ->Chinese (Traditional)
hr.fmuser.org ->Croatian
cs.fmuser.org ->Czech
da.fmuser.org ->Danish
nl.fmuser.org ->Dutch
et.fmuser.org ->Estonian
tl.fmuser.org ->Filipino
fi.fmuser.org ->Finnish
fr.fmuser.org ->French
gl.fmuser.org ->Galician
ka.fmuser.org ->Georgian
de.fmuser.org ->German
el.fmuser.org ->Greek
ht.fmuser.org ->Haitian Creole
iw.fmuser.org ->Hebrew
hi.fmuser.org ->Hindi
hu.fmuser.org ->Hungarian
is.fmuser.org ->Icelandic
id.fmuser.org ->Indonesian
ga.fmuser.org ->Irish
it.fmuser.org ->Italian
ja.fmuser.org ->Japanese
ko.fmuser.org ->Korean
lv.fmuser.org ->Latvian
lt.fmuser.org ->Lithuanian
mk.fmuser.org ->Macedonian
ms.fmuser.org ->Malay
mt.fmuser.org ->Maltese
no.fmuser.org ->Norwegian
fa.fmuser.org ->Persian
pl.fmuser.org ->Polish
pt.fmuser.org ->Portuguese
ro.fmuser.org ->Romanian
ru.fmuser.org ->Russian
sr.fmuser.org ->Serbian
sk.fmuser.org ->Slovak
sl.fmuser.org ->Slovenian
es.fmuser.org ->Spanish
sw.fmuser.org ->Swahili
sv.fmuser.org ->Swedish
th.fmuser.org ->Thai
tr.fmuser.org ->Turkish
uk.fmuser.org ->Ukrainian
ur.fmuser.org ->Urdu
vi.fmuser.org ->Vietnamese
cy.fmuser.org ->Welsh
yi.fmuser.org ->Yiddish
(1) Redundant information of video signal
Taking the YUV component format of recording digital video as an example, YUV represents brightness and two color difference signals respectively. For example, for existing pal TV system, the sampling frequency of luminance signal is 13.5mhz; the frequency band of chroma signal is usually half or less of the brightness signal, which is 6.75mhz or 3.375mhz. Taking the sampling frequency of 4:2:2 as an example, Y signal adopts 13.5mhz, chroma signal U and V are sampled by 6.75mhz, and the sampling signal is quantized by 8bit, then the code rate of digital video can be calculated as follows:
13.5*8 + 6.75*8 + 6.75*8= 216Mbit/s
If such a large amount of data is stored or transmitted directly, it will be difficult to use compression technology to reduce the bit rate. The digital video signal can be compressed according to two basic conditions:
L. data redundancy. For example, spatial redundancy, time redundancy, structure redundancy, information entropy redundancy, etc., that is, there is a strong correlation between pixels of the image. Eliminating these redundancy does not lead to information loss, and it is lossless compression.
L. visual redundancy. Some characteristics of human eyes, such as brightness discrimination threshold, visual threshold, are different in sensitivity to brightness and chroma, which makes it impossible to introduce appropriate errors in coding and will not be detected. The visual characteristics of human eyes can be used to exchange for data compression with certain objective distortion. This compression is lossy.
The compression of digital video signal is based on the above two conditions, which makes the video data greatly compressed, which is conducive to transmission and storage. The common methods of digital video compression are mixed coding, which is to combine transform coding, motion estimation and motion compensation, and entropy coding to compress coding. Usually, transform coding is used to eliminate the intra frame redundancy of the image, and motion estimation and motion compensation are used to remove the inter frame redundancy of the image, and entropy coding is used to further improve the compression efficiency. The following three compression coding methods are introduced briefly.
(a) Compression coding method
(b) Transform coding
The function of transform coding is to transform the image signal described in the space domain to the frequency domain, and then encode the transformed coefficients. Generally speaking, the image has strong correlation in space, and the transformation to frequency domain can realize decorrelation and energy concentration. The common orthogonal transform includes discrete Fourier transform, discrete cosine transform and so on. Discrete cosine transform is widely used in digital video compression.
Discrete cosine transform is referred to as DCT transform. It can transform the image block of L * l from space domain to frequency domain. Therefore, in the process of image compression and coding based on DCT, the image needs to be divided into non overlapping image blocks. Suppose the size of an image is 1280 * 720, it is divided into 160 * 90 image blocks with 8 * 8 size without overlapping in the form of grid. Then DCT transformation can be performed for each image block.
After the block is divided, each 8 * 8 point image block is sent to DCT encoder, and the 8 * 8 image block is transformed from the spatial domain to the frequency domain. The figure below shows an example of an image block of 8 * 8 in which the number represents the brightness value of each pixel. It can be seen from the figure that the brightness values of each pixel in this image block are relatively uniform, especially the brightness value of adjacent pixels is not very large, which indicates that the image signal has a strong correlation.
An actual 8 * 8 image block
The following figure shows the results of DCT transformation of the image block in the above figure. It can be seen from the figure that after DCT transformation, the low frequency coefficient in the upper left corner concentrates a lot of energy, while the energy on the high frequency coefficient in the lower right corner is very small.
The coefficients of image block after DCT transformation
The signal needs to be quantified after DCT transformation. Because human eyes are sensitive to low frequency characteristics of images, such as the overall brightness of objects, and not to the high-frequency details in the image, so in the transmission process, high-frequency information can be transmitted less or not, only the low-frequency part. The quantization process reduces the information transmission by quantifying the coefficients of low frequency region and coarse quantization of the coefficients in high frequency region, which removes the high-frequency information which is not sensitive to human eyes. Therefore, quantization is a lossy compression process and the main reason for the quality damage in video compression coding.
The quantification process can be expressed by the following formula:
Among them, FQ (U, V) represents the DCT coefficient after quantization; f (U, V) represents DCT coefficient before quantization; Q (U, V) represents quantization weighting matrix; q is quantization step; round refers to consolidation, and the value to be output is taken as the closest integer value.
Select the quantization coefficient reasonably, and the result after quantizing the transformed image block is shown in the figure.
DCT coefficient after quantification
Most of DCT coefficients are changed to 0 after quantization, while only a few coefficients are non-zero values. At this time, only these non-zero values need to be compressed and encoded.
(b) Entropy coding
Entropy coding is named because the average code length after coding is close to the entropy value of source. Entropy coding is implemented by VLC (variable length coding). The basic principle is to give short code to the symbol with high probability in the source, and to give long code to the symbol with small probability of occurrence, so as to obtain the shorter average code length statistically. Variable length coding usually includes Hoffman code, arithmetic code, run code, etc. Run length coding is a very simple compression method, its compression efficiency is not high, but the coding and decoding speed is fast, and it is still widely used, especially after the transformation of the encoding, using run-length coding, has a good effect.
First, the AC coefficient immediately following the output DC coefficient of the quantizer shall be scanned in Z-type (as shown in the arrow line). The Z-scan transforms the two-dimensional quantization coefficient into one-dimensional sequence, and then carries on the run length coding. Finally, another variable length code is used to encode the data after the run encoding, such as Hoffman coding. Through this kind of variable length coding, the efficiency of coding is further improved.
(c) Motion estimation and motion compensation
Motion estimation and motion compensation are effective methods to eliminate the correlation of time direction of image sequences. The DCT transform, quantization and entropy coding methods described above are based on one frame image. Through these methods, the spatial correlation between pixels in the image can be eliminated. In fact, in addition to spatial correlation, image signal has temporal correlation. For example, for digital video with background static like news co broadcasting and small movement of main body of picture, the difference between each picture is very small, and the correlation between images is very big. In this case, we do not need to encode each frame image separately, but can only encode the changed parts of adjacent video frames, so as to further reduce the data amount. This work is realized by motion estimation and motion compensation.
Motion estimation technology generally divides the current input image into several small image sub blocks which do not overlap each other, for example, the size of a frame image is 1280 * 720. Firstly, it is divided into 40 * 45 image blocks with 16 * 16 size that do not overlap each other in the form of grid, and then, within the scope of a search window of the previous image or the latter image, find a block for each image block to find one image block within the scope of a search window The most similar image block. The search process is called motion estimation. By calculating the position information between the most similar image block and the image block, a motion vector can be obtained. In this way, the current image block can be subtracted from the most similar image block pointed by the reference image motion vector, and a residual image block can be obtained. Because each pixel value in the residual image block is very small, a higher compression ratio can be obtained in compression coding. This subtraction process is called motion compensation.
Because reference image is needed to be used for motion estimation and motion compensation in the coding process, it is very important to select reference image. Generally, the encoder divides each frame image input into three different types according to the different reference images: I (intra) frame, B (guidance prediction) frame and P (prediction) frame. As shown in the figure.
Typical I, B, P frame structure sequence
As shown in the figure, I frame only uses the data in the frame for coding, and it does not need motion estimation and motion compensation during the coding process. Obviously, since I frame does not eliminate the correlation of time direction, the compression ratio is relatively low. In the process of coding, P frame uses a front I frame or P frame as the reference image for motion compensation, in fact, it encodes the difference between the current image and the reference image. The encoding mode of B frame is similar to P frame, the only difference is that it needs to use a front I frame or P frame and a later I frame or P frame to predict during the coding process. Thus, each P frame coding needs to use one frame image as the reference image, while frame B needs two frames as reference. In contrast, B frame has a higher compression ratio than P frame.
(d) Mixed coding
The paper introduces several important methods in video compression and coding. In practical application, these methods are not separated, and they are usually combined to achieve the best compression effect. The following figure shows the model of hybrid coding (i.e. transform coding + motion estimation and motion compensation + entropy coding). The model is widely used in MPEG1, MPEG2, H.264 and other standards.From the figure, we can see that the current input image must be divided into blocks first, the block of the image obtained by the block shall be subtracted from the predicted image after motion compensation to obtain the difference image x, and then DCT transformation and quantization are carried out for the difference image block. The quantized output data has two different places: one is to send it to the entropy encoder for coding, and the encoded code stream is output to a cache Save in the device and wait for transmission. Another application is to counter quantify and reverse change to signal x ', which adds the image block output with motion compensation to obtain a new prediction image signal, and sends a new prediction image block to frame memory.
|
Enter email to get a surprise
es.fmuser.org
it.fmuser.org
fr.fmuser.org
de.fmuser.org
af.fmuser.org ->Afrikaans
sq.fmuser.org ->Albanian
ar.fmuser.org ->Arabic
hy.fmuser.org ->Armenian
az.fmuser.org ->Azerbaijani
eu.fmuser.org ->Basque
be.fmuser.org ->Belarusian
bg.fmuser.org ->Bulgarian
ca.fmuser.org ->Catalan
zh-CN.fmuser.org ->Chinese (Simplified)
zh-TW.fmuser.org ->Chinese (Traditional)
hr.fmuser.org ->Croatian
cs.fmuser.org ->Czech
da.fmuser.org ->Danish
nl.fmuser.org ->Dutch
et.fmuser.org ->Estonian
tl.fmuser.org ->Filipino
fi.fmuser.org ->Finnish
fr.fmuser.org ->French
gl.fmuser.org ->Galician
ka.fmuser.org ->Georgian
de.fmuser.org ->German
el.fmuser.org ->Greek
ht.fmuser.org ->Haitian Creole
iw.fmuser.org ->Hebrew
hi.fmuser.org ->Hindi
hu.fmuser.org ->Hungarian
is.fmuser.org ->Icelandic
id.fmuser.org ->Indonesian
ga.fmuser.org ->Irish
it.fmuser.org ->Italian
ja.fmuser.org ->Japanese
ko.fmuser.org ->Korean
lv.fmuser.org ->Latvian
lt.fmuser.org ->Lithuanian
mk.fmuser.org ->Macedonian
ms.fmuser.org ->Malay
mt.fmuser.org ->Maltese
no.fmuser.org ->Norwegian
fa.fmuser.org ->Persian
pl.fmuser.org ->Polish
pt.fmuser.org ->Portuguese
ro.fmuser.org ->Romanian
ru.fmuser.org ->Russian
sr.fmuser.org ->Serbian
sk.fmuser.org ->Slovak
sl.fmuser.org ->Slovenian
es.fmuser.org ->Spanish
sw.fmuser.org ->Swahili
sv.fmuser.org ->Swedish
th.fmuser.org ->Thai
tr.fmuser.org ->Turkish
uk.fmuser.org ->Ukrainian
ur.fmuser.org ->Urdu
vi.fmuser.org ->Vietnamese
cy.fmuser.org ->Welsh
yi.fmuser.org ->Yiddish
FMUSER Wirless Transmit Video And Audio More Easier !
Contact
Address:
No.305 Room HuiLan Building No.273 Huanpu Road Guangzhou China 510620
Categories
Newsletter