Method for scalable video coding, device of scalable video coding, software of scalable video coding and machine-readable record medium that saves software

FIELD: information technologies.

SUBSTANCE: share of cast combinations of optimal forecasting modes, which shall be selected for spatially corresponding units of upper and lower layers is identified on the basis of the optimal forecasting mode, which was selected in process of traditional coding, and a table of compliance is developed, which describes interconnections between them. Combinations of selected optimal forecasting modes in the compliance table are narrowed on the basis of the value of the share of casts, in order to create information of compliance for forecasting modes, which describes combinations of narrowed optimal forecasting modes. In process of upper layer unit coding, the version of searching for the forecasting mode, searching for which shall be carried out in process of coding, is identified by referral to information of compliance for forecasting modes using as the key the optimal forecasting mode selected in process of coding of the spatially corresponding unit of the lower layer.

EFFECT: reduced versions of searching for a forecasting mode of an upper layer using correlations of optimal forecasting modes between layers.

7 cl, 14 dwg

 

The technical field

The present invention relates to a method and apparatus for scalable video coding, which is a scalable way to encode video, and the program scalable video coding, which is used to implement this method for scalable video encoding and machine-readable recording medium that stores the program. In particular, the present invention relates to a method and apparatus for scalable video coding, which achieve reduction in encoding time, and the program scalable video coding, which is used to implement this method for scalable video encoding and machine-readable recording medium in which the program is stored.

This application claims the priority of patent applications (Japan) room 2008-271513 registered on October 22, 2008, the contents of which is contained in this document by reference.

The level of technology

In recent years, developed various display terminals and the network environment. As a consequence, the JVT (joint group video standards) performs research system coding-based SVC (scalable video coding), which provides the scalability of the space/time/SNR (signal-to-noise ratio) for AVC (wave chenstohova video coding) (see, for example, non-patent document 1).

In SVC, there are three predictor, namely, mutual forecasting, internal prediction and inter-layer prediction, and is eliminating redundancy inserted between time, space and layers. Examples of prediction modes, reachable through the SVC below.

Mutual forecasting

- skip mode (Skip)

- direct mode (Direct)

- prediction mode motion with a block size of 16×16 (P16×16)

- prediction mode motion with a block size of 16×8 (P16×8)

- prediction mode motion with a block size of 8×16 (P8×16)

- prediction mode motion with a block size of 8×8 (P8×8)

Internal forecasting

mode internal prediction block size 16×16 (I16×16)

mode internal prediction block size 8×8 (I8×8)

mode internal prediction block size 4×4 (I4×4)

Interlayer prediction

the BLSkip mode (BLSkip)

mode IntraBL (IntraBL)

When P8×8, each block of 8×8 can be further divided into block sizes of 8×4, 4×4 and 4×4. In SVC one of these options for finding prediction mode (search mode prediction) is selected as the optimal prediction mode of each macroblock.

An example of how, using the constituent for to determine the optimal prediction mode, below.

In JSVM (joint scalable video model: see, for example, non-patent document 2), which moves JVT as support SVC-encoder, the cost of encoding, which are formed from bits of coding and distortion when encoding, are calculated in each prediction mode and the prediction mode having the smallest cost on the coding of all of the above modes forecasting is defined as the optimal prediction mode.

In addition, in patent document 1 below, the vectors are generated through extrapolation or interpolation of motion vectors based on the reference frame for encoding the present frame, and the coordinates of each pixel in the macroblock, which are displaced as a result of this, then determined, and the number of times the pixels are the same, is calculated for each pixel. Then, searches the prediction mode are narrowed in accordance with the value of the measure, which is calculated from the counted number for each pixel within the target macroblock encoding. The method used for this narrow is the way which is suggested in order to increase the search speed prediction mode in H.264/AVC, however, it can also be used in SVC which is the mechanism for finding prediction mode, identical to the search mode prediction in H.264/AVC.

In addition, in patent document 2, below, to allow you to perform intra-frame coding at high speed, for example, nine of the display forecast errors are defined in the block in which intra-frame coding must be performed using the pixel values of the adjacent blocks of the encoding, and, on the basis of these forecast errors, the prediction mode is defined for this block. Then, the prediction mode for the block is determined using the on-screen prediction mode of the adjacent block that has already been encoded, and when these two prediction mode are the same, the prediction mode is selected under such conditions. If, however, these two modes of prediction do not match, the selected prediction mode having lower costs for encoding.

The documents of the prior art,

Non-patent document 1. T. Wiegand, G. Sullivan, J. Reichel, H. Schwarz and M. Wien: "Joint Draft ITU-T Rec. H.264-ISO/IEC 14496-10/ Amd.3 Scalable video coding, "ISO/IEC JTC1/SC29/WG11 and ITU-T SG16 Q.6, JVT-X201, 2007. http://ftp3.itu.ch/av-arch/jvt-site/2007_06_Geneva/JVTX201.zip

Non-patent document 2. J. Reichel, H. Schwarz and M. Wien: "Joint Scalable Video Model JSVM-11", ISO/IEC JTC1/SC29/WG11 and ITU-T SG16 Q.6, JVT-X202, 2007. http://ftp3.itu.ch/av-arch/jvt-site/2007_06_Geneva/JVTX202.zip

Patent document 1. Has not passed the examination of the patent application (Japan), first publ who used (JP-A) No. 2006-033451

Patent document 2. Has not passed the examination of the patent application (Japan), first publication (JP-A) No. 2005-184241

The invention

Problem resolved by the invention of

In JSVM-method of determining the optimal prediction mode, described in non-patent document 2, because there is no narrowing of the search mode prediction can be achieved superior performance coding. On the other hand, in the method of determining an enormous amount of time required to perform the search mode prediction. Namely, in this method of determining, for a long time have been wasted, since it searches for all modes of prediction even for forecasting, which, obviously, have a low probability of choice, when considering the characteristics of the images within the macroblock (e.g., modes of forecasting in internal mode in static fields).

In addition, since the narrowing of the search mode prediction in patent document 1 is a method for determining whether to perform or not forecast in the internal mode, it does not influence the reduction of the search mode of mutual prediction, which require longer calculation time compared to searching for the internal mode is prognozirovaniya. Namely, significant room for improvement remains in relation to searches of mutual prediction.

In addition, since the narrowing of the search of the prediction mode according to patent document 2 is only a contraction for internal forecasting, similar to the narrowing of the search of the prediction mode according to patent document 2, there is no impact on reducing search mode mutual prediction. Namely, significant room for improvement remains in relation to searches of mutual prediction.

The present invention was created in consideration of the above circumstances, and its purpose is to provide a new technology for scalable video coding, which, when scalable video encoding that achieves scalability through multi-layered structure narrows the search options of the prediction mode of the upper layer using the correlations of the optimal prediction modes between layers in order to increase the speed.

The solution to the problem

When scalable video encoding, in which scalability is achieved through multi-layer structure, in order to increase the search speed of the prediction mode, the device is scalable video coding in the present invention includes: (1) generation module, which, on the basis of information about the optimal prediction mode that is selected when scalable coding is performed without imposing restrictions on the use of prediction modes specified as possible to use, determines the share of depositions combinations of optimum prediction modes, which should be selected for spatially corresponding blocks of the upper layer and the lower layer, and creates a lookup table that describes the relationship between the combinations of the selected optimum prediction mode and the optimal prediction mode that should be selected, and shares of depositions; (2) a detection module, which, when encoding a block of the upper layer, detects information about the optimal prediction mode, selected when the spatial encoding of the corresponding block of the lower layer; (3) determining module that, based on the information about the selected optimum prediction mode detected by the detection module, and on the basis of information about the shares of the depositions described in the table of correspondence, extracts the effective combination from among combinations described in the lookup table, and determines the optimum prediction mode of the upper layer contained in the extracted effective combination, as an option to search the tunes forecasting, the search must be performed when encoding a block of the upper layer; and (4) a control module that controls so that scalable coding, in which restrictions are imposed on the use of prediction performed using a lookup table, and scalable coding, in which restrictions are not imposed on the use of prediction modes to be performed without using the lookup table are repeated alternately.

In such a structure as described above, by reference to the lookup table, using as key information about the optimal prediction modes, which are detected by the detection module, the determining module indicates the proportion of the precipitation associated with these optimal prediction. Then, preferably for a module definition or to extract a combination of optimum prediction modes, with the share of precipitation, showing a value greater than a predetermined threshold value from among those specified share of the deposition, or to extract a combination of optimum prediction modes, with the share of precipitation, showing the highest value of these listed shares depositions, or to extract the optimal combination of modes prognosis is, having a predetermined number of shares of depositions, which are selected sequentially from the proportion of precipitation that has the largest value from among those specified fractional precipitation. The determining module then determines the optimum prediction mode of the upper layer contained in the extracted combination of optimum prediction modes, as search variant of the prediction mode, the search must be performed when encoding a block of the upper layer.

In addition, to make the treatment effective determination by the determining module, it is preferable for the device scalable video encoding of the present invention to extract in advance of the effective combination of optimum prediction modes by narrowing combinations of optimum prediction modes described in the table, based on the values of fractional precipitation, described in the table, and create information compliance for forecasting, which describes the combination of the extracted effective optimal prediction modes.

In this case, when the scalable video encoding, in which scalability is achieved through multi-layer structure, in order to increase the search speed of the prediction mode, the device masstabi is imago video encoding of the present invention includes: (1) a module to create lookup tables, which, on the basis of information about the optimal prediction mode that is selected when scalable coding is performed without imposing restrictions on the use of prediction modes specified as possible to use, determines the share of depositions combinations of optimum prediction modes, which should be selected for spatially corresponding blocks of the upper layer and the lower layer, and creates a lookup table that describes the relationship between the combinations of the selected optimum prediction mode and the optimal prediction mode that should be selected, and shares of depositions; (2) creation module information compliance for forecasting, which is based on the values of fractional precipitation, extracts a combination of efficient optimal prediction modes by narrowing combinations of the selected optimum prediction modes described in the table, and the optimal prediction mode that should be selected, and creates information compliance for forecasting, which describes the combination of the extracted effective optimal prediction modes; (3) a detection module, which, when encoding a block of the upper layer, detects information about the optimal prediction mode, wybran the m spatial encoding of the corresponding block of the lower layer; (4) the module definition, which, by reference to the information compliance for forecasting using as key information about the selected optimum prediction mode detected by the detection module, is determined by the search variant of the prediction mode, the search must be performed when encoding a block of the upper layer; and (5) a control module that controls so that scalable coding, in which restrictions are imposed on the use of prediction performed using a lookup table, and scalable coding, in which restrictions are not imposed on the use of prediction modes to be performed without using the lookup table repeated alternately.

When this structure is used, the generation module information compliance for forecasting can either: create information compliance for prediction by extracting, as an effective combination, a combination of optimum prediction modes having a share of depositions, showing the value that exceeds a predetermined threshold value; or, alternatively, to create information compliance for prediction by extracting, in which the quality of an effective combination, the combination of optimum prediction modes having a share of depositions, showing the highest value, from among combinations of optimum prediction modes that have identical optimal prediction mode on the lower layer; or create information compliance for prediction by extracting, as an effective combination, a combination of a predetermined number of optimal prediction modes selected sequentially from the proportion of precipitation, showing the highest value.

Method for scalable video encoding of the present invention, which is implemented as a result of each of the above processing modules may also be implemented through computer program. This computer program provides the possibility of implementing the present invention by suitable machine-readable recording media, or by providing over the network or by installing on the computer and control via the control module, such as a CPU, when the present invention should be implemented.

Advantages of the invention

In the present invention, a scalable video encoding that achieves scalability through multi-layer structure, since it is possible to narrow the search option of the prediction mode of the upper layer using the correlations of the optimal prediction modes between layers, reducing the encoding time can be achieved.

In addition, in the present invention, when the reduction time encoding is achieved by narrowing the search mode prediction, because this narrowing is performed based on the corresponding relationships of optimum prediction modes between layers in the already-encoded frames, you can eliminate the risk of missing a search of the optimal prediction mode as a result of this narrowing. Accordingly, it is possible to suppress the degradation of coding, which may result from narrowing of the search mode prediction.

Brief description of drawings

1 is an explanatory view showing examples of frames intended for encoding, and personnel designed to create a table of the degrees of match of the prediction modes.

Figure 2 is a block diagram of the operational sequence of the method, showing a generalized sequence of processing operations of the video encoding according to a variant implementation of the present invention.

Figure 3 is a table of the degrees of match of the prediction modes according to a variant implementation of the present invention.

Figure 4 is a table showing the results when the search mode, the forecast is of the taper in the embodiment of the present invention.

Figure 5 is a table showing the results when the search mode prediction taper in the embodiment of the present invention.

6 is a flowchart of the operational sequence of the method showing the processing of scalable video coding according to a variant implementation of the present invention.

7 is a flowchart of the operational sequence of the method, showing an example of processing in order to identify options for the search of the prediction mode, which is performed in the processing of the scalable video encoding, shown in Fig.6.

Fig is a flowchart of the operational sequence of the method, showing another example of processing to determine the search options of the prediction mode, which is performed in the processing of the scalable video encoding, shown in Fig.6.

Fig.9 is a block diagram showing the device of scalable video coding according to a variant implementation of the present invention.

Figure 10 is a block diagram showing an example of a module definition options search mode prediction unit scalable video coding, is shown in Fig.9.

11 is a block diagram showing another example of a module definition options search prediction mode in which trojstva scalable video coding, shown in Fig.9.

Fig is an explanatory view showing the frame of the calculation of compliance and frame high-speed mode is selected in the experiment, which is designed to test the effectiveness of a variant of implementation of the present invention.

Fig is a graph showing experimental results of the experiment performed to test the effectiveness of a variant of implementation of the present invention.

Fig is a graph showing experimental results of the experiment performed to test the effectiveness of a variant of implementation of the present invention.

The implementation of the invention

[1] the Basic idea of embodiments of the present invention

In the embodiment of the present invention, a scalable video encoding that achieves scalability through multi-layered structure, increasing the speed at which the search is performed of the prediction mode, is achieved by two processes, namely:

(i) by creating a table of the degrees of match modes forecasting (i.e. the matrix that describes the correlation of optimum prediction modes between layers); and

(ii) by narrowing the search mode prediction using this table of degrees according to tsti modes of prediction.

Further in this document, the description proceeds in accordance with the example shown in figure 1. Namely, it is assumed that as the layer L and layer L-1 is encoded using IBBBP-ordered B-patterns. The arrows in the diagram show the basic purpose of forecasting. The target layer coding is considered as L, the target frame coding is considered as B2b, and the frame intended to create a table of the degrees of match modes forecasting is considered as B2a. In addition, the frame layer L-1 are identical synchronization with B2b is considered as a B 2b, and the frame layer L-1 are identical synchronization B2a is considered as a B 2a. The coding is performed sequentially from the lowest temporal level, and within the same time level, the coding is performed sequentially from the frame with the earliest time. The layers are encoded in this sequence from the lowest level.

Next, with respect to the flowchart of the operational sequence of the method shown in figure 2, describes a generalized sequence of processing operations of this variant implementation.

In the present embodiment, when scalable coding is performed for the video, as shown in the block diagram of the operational sequence of the method in figure 2, the variable n is set equal to 1 at the step S101, and you is within each group definition, all or no frames encoded, the next step S102. If it is determined that all the frames are encoded, the procedure ends.

If, however, in accordance with the processing determination step S102 is determined that not all frames are encoded, the procedure goes to step S103, where one raw frame is selected in accordance with the sequence from the initial frame. In the next step S104, the selected frame is encoded through the implementation of the forecast without imposing restrictions on the use of prediction modes are specified as available for use, namely through the implementation of a forecast using all applicable modes of prediction.

Then, at step S105, the value of the variable n is incremented. In the next step S106, the system determines whether exceeds or not the value of the variable n pre-defined threshold value N1 (N1 is an integer equal to 1 or greater). If it is determined that the value of the variable n exceeds the threshold value N1, the procedure returns to the processing of step S102, and the encoding frame continues without imposing restrictions on the use of prediction modes are specified as approved for use.

If, however, the processing determination step S106, opredelyaetsya, what is the value of the variable n exceeds the threshold value N1, the procedure goes to step S107, and creates a table of the degrees of match modes forecasting. Table of degrees of match modes forecasting has such a data structure, as described below, is a table that describes the correlation (i.e. the share of depositions) optimal prediction modes between layers.

Then, at step S108, the variable n is set equal to 1, and, in the next step S109, the system determines whether or not all frames are encoded. If it is determined that all the frames are encoded, the procedure ends.

If, however, at step S109 is determined that not all frames are encoded, the procedure goes to step S110. At step S110, one raw frame is selected in accordance with the sequence from the initial frame. In the next step S111, the selected frame is encoded through the implementation of the forecast while narrowing the search mode prediction using the table of degrees of match of the prediction modes.

Then, at step S112 (the value of the variable n is incremented. In the next step S113, the system determines whether exceeds or not the value of the variable n pre-defined threshold value N2 (N1 is an integer, R is ate 1 or more). If it is determined that the value of the variable n exceeds the threshold value N2, the procedure returns to the processing of step S109, and the encoding frame continues while narrowing the search mode prediction using the table of degrees of match of the prediction modes.

If, however, at the step S113 is determined that the value of the variable n exceeds the threshold value N2, it is determined the need to update the table of degrees of match of the prediction modes, and the procedure returns to step S101. Accordingly, the processing of steps S101-S113 continues in the time table of the degrees of match modes forecasting is updated.

Thus, in the present embodiment, when scalable coding is performed for the video, after the N1-th frame is encoded based on the results of this coding creates a table of the degrees of match of the prediction modes, which describes the correlation (i.e. the share of depositions) optimal prediction modes between layers. Then, the procedure proceeds to the subsequent encoding N2-frame, and the processing is repeated to encode the N2-th frame while narrowing the search mode prediction using the table of degrees of match modes prognosis the project.

(i) create a table of the degrees of match modes forecasting

The following describes the processing of creating the table of degrees of match modes forecasting performed at step S107.

In the processing of step S104, for frame B2a created using the table of degrees of match of the prediction modes, and frame B 2a directly below it, as shown in figure 1, the encoding has been completed, and the optimal prediction modes is selected. When frames B2a and B 2a encoded information about the selected optimum prediction mode is stored in the buffer. Based on this information, the optimal prediction mode, which is stored in the buffer, analyzes the relationship of correspondence between the optimal prediction mode for a macroblock (hereinafter in this document abbreviated as MB) frame B2a and submicromolar (hereinafter in this document abbreviated as SMB) space corresponding frame B 2a.

In particular, the table of the degrees of match of the prediction modes having such a data structure, as shown in Figure 3 is created between B2a and B 2a. The numeric values shown in Figure 3 show the ratio (i.e. the proportion of precipitation), in which, when the optimal prediction mode selected in each SMB (size 8×8) of the frame B 2a, is i, the optimal mode j forecasting of you who ireda in MB frame B2a. For example, if P16×16 is selected in the SMB frame B 2a, MB frame B2a, which spatially corresponds to the SMB, in which the selected P16×16, it shows that selects the skip mode 32,3%.

Here, the method used to choose the optimal prediction mode in frames B2a and B 2a, can be JSVM-the method described in non-patent document 2, or the way in which searches of the prediction mode is narrowed, for example, the method described in patent document 1.

In addition, in this example, the target frame table of the degrees of match modes forecasting is considered as one frame (i.e. B2a), which is already coded and has a temporal level, identical to the temporal level of the target frame, which must be encoded, however, is not limited to the aforesaid. You can also choose the target frame, which is already coded and has a different time level (e.g., B1). In addition, you can also target multiple frames (e.g., B1 and B2a) and calculate the degree of match using the sum of this set of frames. Namely, assuming that the encoding of the frames have already been completed in the target layer coding and in the layer directly below it, then they can be the target frames, which should be generated by the table of degrees of match modes forecasting.

(ii) Narrowing the search mode prediction using the table of degrees of match modes forecasting

Below is a description of processing to narrow down searches of prediction mode using the table of degrees of match of the prediction modes, described in step S111.

Search mode prediction taper in each MB of the target frame B2b encoding in accordance with the values of the degrees of match of the prediction modes in the table of degrees of match of the prediction modes that you created in step S107. The numerical values in the table of degrees of match of the prediction modes show the probability of finding the optimal prediction mode in a specific target macroblock encoding.

Further more describes processing in order to narrow search mode prediction. In the description below, the target macroblock coding frame B2b written as MBL, while submicromolar frame B 2b layer L-1, which is the spatial position identical to the position of this MBL recorded as SMBL-1.

When narrowing options search prediction mode macroblock MBL, first, read the information about the optimal prediction mode of submicromolar SMBL-1. Then, the table of degrees of match modes cons the licensing mapped to the optimal prediction mode SMBL-1, and gather information on the probability (i.e. the fraction of spins that each prediction mode in the target macroblock MBL encoding mode is best prediction. Then, on the basis of this probability is the optimal prediction mode, search mode prediction narrowed. Two examples of this narrowing below.

(a) the Procedure of narrowing 1

The procedure of narrowing 1 is a procedure for narrowing the search mode prediction using threshold values for narrowing the search of the prediction mode.

In this procedure, narrowing to 1, the threshold value t% for narrowing the search of the prediction mode is provided, and the prediction modes, which are smaller than the threshold value t%, are excluded from the search. The value of threshold t is provided from the outside. An example of a method used to determine this value, is the way in which a value that limits the performance degradation of the coding part of the allowable range is determined through repeated execution of the processing of the encoding.

Here, if the way in which the probability (i.e. the degree of match) that each prediction mode in the MBL is the optimal prediction mode are considered ivalsa from a table of the degrees of match modes forecasting in time, when the information about the optimal prediction mode for SMBL-1 is detected and compared with a threshold value to narrow search mode prediction, the required processing comparison becomes extremely difficult.

Therefore, the treatment threshold is performed in advance on the degree of the correspondence table of the degrees of match modes forecasting using thresholds narrow search mode prediction so that the degree of correspondence correspondence table of the prediction modes is converted to binary form.

The results of narrowing options search prediction mode, when the threshold value for narrowing the search of the prediction mode set to 5% in the table of degrees of match of the prediction modes shown in Figure 3, is shown in Figure 4. In figure 4, ○ represents search suggestions, and x represents the prediction modes, are excluded from the search.

(b) the Procedure of narrowing 2

The procedure of narrowing 2 is a procedure in which only the prediction modes in which the degree of match of the prediction modes is maximum, can be specified as search suggestions.

The prediction modes in which the degree of match of the prediction modes is maximum, zagaytsy as search suggestions. In this case, search suggestions usually be shrunk to a single prediction mode, however, when there are many options search prediction mode which provides the maximum value, they are set as search suggestions.

Here, if the way in which the probability (i.e. the degree of match) that each prediction mode in the MBL is the optimal prediction mode, is read from the table of degrees of match modes forecasting at the time when the information about the optimal prediction mode for SMBL-1 is detected, and the degree of correspondence with the maximum value from their number, specified the required processing instructions becomes extremely difficult.

Therefore, the degree of correspondence with the maximum value that is contained in the degrees of correspondence correspondence table of the prediction modes, and the degree of correspondence correspondence table of the prediction modes is converted to binary form.

The results of the contraction, when the prediction mode having the maximum value in the table of degrees of match of the prediction modes shown in Figure 3, is set as the search variant of the prediction mode, shown in figure 5. Figure 5, ○ represents search suggestions, and represents the prediction modes, excluded from the search.

Further in this document, the present invention is described in detail in accordance with the variants of the implementation.

6 to 8 are block diagrams of a sequence of operations of the method showing the processing of the scalable video encoding performed by this version of the implementation.

6 is a flowchart of the operational sequence of the method, showing a complete processing of the scalable video encoding performed by this version of the implementation. 7 and Fig are a flowchart of the operational sequence of the method, showing the example and another example of details of processing executed at the step S201 of the flowchart of the operational sequence of the method shown in Fig.6.

Next, the processing of the scalable video encoding performed by this version of the implementation, is described in detail in accordance with these by flowchart the sequence of operations.

Processing coding this option is treatment for improving layers, and processing of non-scalable single-layer coding is applied to the base layer. An example of processing of the single-layer encoding is the processing of the encoding of the base layer reference SVC-encoder JVSM, mentioned the th in non-patent document 2.

The following describes the processing of steps S201-S206 performed in the block diagram of the sequence of operations of the method shown in Fig.6.

At step S201, read the initial value of the search-mode prediction search which must be executed in the target macroblock (MB) coding, and search options for prediction mode, the search of which must be ultimately in the target MB coding, are determined and stored in the register. This processing is described in detail below in relation to Fig.7 and Fig.

At step S202, the information about searches of the prediction mode, which is stored as the processing of step S201, is read from the register, and searches each search variant of the prediction mode. One optimum prediction mode that should be used for coding, is then determined, and information about this mode is stored in the register. One example of the method of solving of the optimal prediction mode is the method that runs in JSVM, in which the prediction mode, which minimizes the cost of encoding, which is expressed by a linear sum of the bit encoding and distortion when encoding, is considered as the optimal prediction mode.

At step S203, the information about the optimal prediction mode in Celje is om MB coding is read from the register, and performs motion compensation in this case, the optimal prediction mode. The residual signal prediction is then created and stored in the buffer.

At step S204, the residual signal prediction is read from the buffer, and this residual signal prediction is then encoded, and the resulting encoded data stored in the buffer. One example of this processing is the processing sequence DCT, quantization and variable length coding in support SVC-JSVM encoder mentioned in non-patent document 2.

At step S205, the processing of determining is performed in order to determine what has been completed or no coding just MB. If the encoding of the entire MB is completed, the processing of coding ends, and the coded data of each MB and other necessary header information is read from the buffer and displayed as the final coded data. If the encoding of the entire MB is not completed, the procedure goes to step S206.

At step S206, the processing proceeds to the next target MB coding, and processing of step S201.

Next, a specific example of processing executed at the step S201 described using a flowchart of the operational sequence of the method shown in Fig.7, which includes the stages of spider S301 demonstration-S306.

Phase spider S301 demonstration, reads information, which indicates t is, whether or not MB, designed for encoding, MB, intended to narrow the search of the prediction mode, which is the purpose of this variant implementation. If MB designed for encoding, MB is designed to narrow down search of the prediction mode, the procedure proceeds to processing of step W302. If MB designed for encoding, MB is not intended to narrow the search of the prediction mode, the initial value of the search variant of the prediction mode is output as a final search of the prediction mode.

On stage, W302, information indicating already-encoded frames, which are used to calculate the table of degrees of match modes forecasting, is read from the outside, and the information of the prediction mode about these frames is stored in the register.

At step S303, reads the information of the prediction mode (i.e. the information about the optimal prediction mode used for encoding) in frames that are designed to calculate the table of degrees of match of the prediction modes, and the degree of match (i.e. the share of depositions) optimal prediction layer, designed for encoding, and the layer directly below it are calculated and stored in the register as the table of degrees of match modes forecasting. Such a table of the degrees of match of the prediction modes, as shown in Figure 3, is created and stored in the register.

At step S304, read the table of degrees of match of the prediction modes, and this table is then stored in the buffer.

At step S305, read threshold narrowing search of the prediction mode and then stored in the register.

At step S306, the table of degrees of match modes forecasting is read from the buffer, and the threshold values of the narrow search mode prediction are also read from the register. Only the prediction modes in which the degree of matching (i.e., the fraction of spins is equal to or exceeds the threshold narrowing search of the prediction mode, specified as search options in the prediction mode, and information about them is stored in the register. When these settings and the save operation, selects, selects and saves only those options search prediction mode associated with the optimal prediction mode obtained by encoding the base layer.

Thus, in the block diagram of the sequence of operations of the method shown in Fig.6, based on a table of the degrees of match of the prediction modes having such a data structure, as shown in Figure 3, however the fast processing for to narrow search mode prediction in this format, as shown in Figure 4.

Next, another specific example of processing executed at the step S201 described using a flowchart of the operational sequence of the method shown in Fig, which includes the steps S401-S405.

At step S401, reads information, which indicates whether or not MB, designed for encoding, MB, intended to narrow the search of the prediction mode, which is the purpose of this variant implementation. If MB designed for encoding, MB is designed to narrow down search of the prediction mode, the procedure proceeds to processing of step S402. If MB designed for encoding, MB is not intended to narrow the search of the prediction mode, the initial value of the search variant of the prediction mode is output as a final search of the prediction mode.

At step S402, the information indicating already-encoded frames, which are used to calculate the table of degrees of match modes forecasting, is read from the outside, and the information of the prediction mode about these frames is stored in the register.

At step S403, reads the information of the prediction mode (i.e. information about Optim is flax prediction mode, used when encoding) in frames that are designed to calculate the table of degrees of match modes forecasting. Then, the degree of matching (i.e. the share of depositions) optimal prediction layer, designed for encoding, and the layer directly below it are calculated and stored in the register as the table of degrees of match modes forecasting. Namely, such a table of the degrees of match of the prediction modes, as shown in Figure 3, is created and stored in the register.

At step S404, read the table of degrees of match of the prediction modes, and this table is then stored in the buffer.

At step S405, the table of degrees of match modes forecasting is read from the buffer, and only the prediction mode having the maximum degree of matching (i.e. the proportion of precipitation), is specified as a search variant of the prediction mode, and information about it is stored in the register. Here, when are these settings and the save operation, selects, selects and saves only those options search prediction mode associated with the optimal prediction mode obtained by encoding the base layer.

Thus, in the block diagram of the sequence of operations of the method shown in Fig.6, table-based degrees with twelve modes forecasting, having such a data structure, as shown in Figure 3, are processed in order to narrow search mode prediction in this format, as shown in Figure 5.

Figures 9-11 show the device structure of scalable video coding according to a variant implementation of the present invention.

Fig.9 shows the device structure of a scalable video encoding this variant implementation. Figure 10 and 11 show an example and another example of a detailed structure of the module 102 identifying options for finding prediction mode shown in Fig.9.

Further, the device scalable video encoding this variant implementation is described in detail in relation to these structural diagrams of the device.

The device is scalable video encoding this variant implementation is a processing device for improving layers, and processing of non-scalable single-layer coding is applied to the base layer. An example of processing of the single-layer encoding is the processing of the encoding of the base layer reference SVC-encoder JVSM mentioned in non-patent document 2.

First, the full structure of the device scalable video coding is described with reference to Fig.9.

Module 101 storing initial values of the search suggestions dir the mA prediction reads the initial value of the search mode prediction, and outputs them to the register.

The module 102 identifying options for finding prediction mode reads the initial values for the search of the prediction mode and defines the search options of the prediction mode, a search which ultimately must be performed. Then, the module 102 identifying options for finding prediction mode displays in the register information about searches of the prediction mode, which is ultimately determined, and processing proceeds to module 103 determination of optimum prediction modes. The detailed structure of this module processing is described below using Figure 10 and 11.

The module 103 to determine the optimal prediction modes reads options search prediction mode from the register and performs a lookup for each search variant of the prediction mode. Then, the module 103 to determine the optimal prediction modes to determine the optimum prediction mode that should be used for encoding, and outputs information about this mode, the module 104 storing the optimal modes of prediction. One example of the method of solving of the optimal prediction mode is the method that runs in JSVM, in which a prediction mode in which the cost of encoding, which is expressed by a linear sum of the bit encoding and travesty the Oia when encoding, minimized, is considered as the optimal prediction mode.

The module 105 to generate the residual signals of the prediction reads the optimum prediction mode in the target MB encoding module 104 storing the optimal prediction modes, and performs motion compensation in this case, the optimal prediction mode. He then creates a residual signal prediction and outputs it to the buffer.

Module 106 encoding residual signals of the prediction then reads the residual signal prediction in MB, designed for encoding, from the buffer and performs the encoding for this residual signal prediction. The coded data is then outputted to the buffer. One example of this processing is the processing sequence DCT, quantization and variable length coding support SVC-encoder JSVM mentioned in non-patent document 2.

Module 107 determine the complete coding MB performs processing of determining to determine what has been completed or no coding just MB. If the encoding of the entire MB is completed, the processing of coding ends, and the final coded data is displayed. If the encoding of the entire MB is not completed, the procedure proceeds to the processing module 108 updates the target MB encoding.

The module 108 updates the target MB encoding transfer which leads to the next MB, designed for encoding, and the processing module 102 identifying options for finding prediction mode.

Next, an example of a detailed structure of the module 102 define search of the prediction mode is described with reference to Figure 10.

Module 201 storing information specifying the target MB narrowing the search mode prediction reads information indicating that whether or not MB be narrowing search mode prediction, and outputs this information to the register.

Module 202 determining the target MB narrowing the search mode prediction reads the information indicating the MB, which should be a narrowing of the search of the prediction mode, the module 201 storing information specifying the target MB narrowing search of the prediction mode, and performs processing in order to determine whether or not MB, designed for encoding, MB, which should narrow. If MB designed for encoding, is MB, which should narrow, the procedure proceeds to the processing module 206 create tables of degrees of match modes forecasting. If MB designed for encoding, not a MB, which should narrow, then the initial value of the search variant of the prediction mode is determined as the target is Ariant search prediction mode and displayed.

Module 203 storing information specifying a target frame calculate the degree of match modes forecasting reads the information indicating the frames that are already coded and designed to calculate the degree of match of the prediction modes, and displays it in the register.

Module 204 storing the optimal modes of forecasting improving layer for the target frame reads the information of the optimum prediction mode of the layer intended for encoding, for frames that are designed to calculate the degree of match of the prediction modes, which are specified by the information specifying read from module 203 storing information specifying a target frame calculate the degree of match of the prediction modes, and displays it in the register.

Module 205 storage for optimal prediction modes for the layer directly under the target frame reads the information of the optimum prediction mode of the layer immediately below the layer intended for the coding for frames that are designed to calculate the degree of match of the prediction modes, which are specified by the information specifying read from module 203 storing information specifying a target frame calculate the degree of match of the prediction modes, and displays it in p is of gestr.

The module 206 create tables of degrees of match modes forecasting reads from the module 204 storing the optimal modes of forecasting improving layer for the target frame information of the optimum prediction mode in the layer, designed to encode a frame designed to calculate the degree of match of the prediction modes. In addition, the module 206 create tables of degrees of match modes forecasting also reads the information of the optimum prediction mode of the layer immediately below the layer intended for the coding of frames that are used to calculate the degree of match modes forecasting module 205 storage for optimal prediction modes for the layer directly under the target frame, and calculates the degree of matching (i.e. the share of depositions) optimal prediction modes between the respective macroblocks and submicroscale and displays the results in the module 207 for storing tables of degrees of match modes forecasting as a table of the degrees of match of the prediction modes.

Module 208 storing threshold values of the narrow search mode prediction reads the threshold value for narrowing the search of the prediction mode, and outputs them to the register.

Module 209 comparison threshold is o the table values of the degrees of match modes forecasting reads the table of degrees of match modes forecasting module 207 for storing tables of degrees of match modes forecasting and also reads the threshold value for narrowing search mode prediction module 208 storing threshold values of the narrow search mode prediction. Then, the module 209 comparison of threshold values table of the degrees of match modes forecasting analyzes the probability of the optimal prediction mode MB designed for encoding, which is associated with the optimal prediction mode SMB directly under him, and then specifies only the prediction modes, the probability of which is equal to or exceeds the threshold narrowing the search mode prediction, as a search of the prediction mode, and outputs them.

Thus, in the structure of the device shown in Figure 10, based on a table of the degrees of match of the prediction modes having such a data structure, as shown in Figure 3, are processed in order to narrow search mode prediction in this format, as shown in Figure 4.

Next, another example of a detailed structure of the module 102 define search of the prediction mode is described with reference to 11.

Module 301 storing information specifying the target MB narrowing the search mode prediction reads the information indicating the need or h is t MB be narrowing search of the prediction mode, and displays this information in the registry.

Module 302 determine the target MB narrowing the search mode prediction reads the information indicating the MB, which should be a narrowing of the search of the prediction mode, the module 301 storing information specifying the target MB narrowing search of the prediction mode, and performs processing in order to determine whether or not MB, designed for encoding, MB, which should narrow. If MB designed for encoding, is MB, which should narrow, the procedure proceeds to the processing module 306 create tables of degrees of match modes forecasting. If MB designed for encoding, not a MB, which should narrow, then the initial value of the search variant of the prediction mode is determined as the final search of the prediction mode and displayed.

Module 303 storing information specifying a target frame calculate the degree of match modes forecasting reads the information indicating the frames that are already coded and designed to calculate the degree of match of the prediction modes, and displays it in the register.

Module 304 storing the optimal modes of forecasting improving layer for the target frame reads the information of the optimal mode about what noshirvani layer, designed for encoding, for frames that are designed to calculate the degree of match of the prediction modes, which are specified by the information specifying read from the module 303 storing information specifying a target frame calculate the degree of match of the prediction modes, and displays it in the register.

Module 305 storage for optimal prediction modes for the layer directly under the target frame reads the information of the optimum prediction mode of the layer immediately below the layer intended for the coding for frames that are designed to calculate the degree of match of the prediction modes, which are specified by the information specifying read from the module 303 storing information specifying a target frame calculate the degree of match of the prediction modes, and displays it in the register.

The module 306 create tables of degrees of match modes forecasting reads from the module 304 storing the optimal modes of forecasting improving layer for the target frame information of the optimum prediction mode in the layer, designed to encode a frame designed to calculate the degree of match of the prediction modes. In addition, the module 306 create tables of degrees of match modes prognose the Finance also reads the information of the optimum prediction mode of the layer immediately below the layer, designed for encoding frames that are used to calculate the degree of match modes forecasting module 305 storage for optimal prediction modes for the layer directly under the target frame. Then, the module 306 create tables of degrees of match modes forecasting calculates the degree of matching (i.e. the share of depositions) optimal prediction modes between the respective macroblocks and submicroscale and displays the results in the module 307 for storing tables of degrees of match modes forecasting as a table of the degrees of match of the prediction modes.

The module 308 analysis of the prediction mode to the maximum share of depositions reads the table of degrees of match modes forecasting module 207 for storing tables of degrees of match modes forecasting and analyzes the probability of the optimal prediction mode MB designed for encoding, for optimal prediction mode SMB directly under him, and then sets the prediction mode, the probability of which is maximum, as a final search of the prediction mode and outputs it.

Thus, in the structure of the device shown figure 11, based on the table of the degrees of match modes prognose the Finance, having such a data structure, as shown in Figure 3, are processed in order to narrow search mode prediction in this format, as shown in Figure 5.

Next, we describe the results of an experiment which is performed to verify the effectiveness of the present invention.

This experiment was performed by setting this option implementation in JSVM 9.12.2 and the subsequent comparison of JSVM with the present embodiment. The images are test images SVC "City" and "Football"with the size of 704×576, and "Pedestrian" and "Station"with a size of 1920×1024. The image with the above resolution was entered in improving layer, and the image having the resolution in number of pixels vertically and horizontally equal to half of the above, introduced into the base layer. The number of frames of the encoded image is 129, and four QP (quantization parameter), namely, 22, 27, 32 and 37, tested using the same values on both layers. The structure of a GOP (group of pictures) is IBBBP-hierarchical structure of B-pictures, and I entered every 16 frames. As shown in Fig, two frames belonging to the lowest temporal level, used, respectively, for the frame calculate the degree of match levels and high-speed frame the selection mode. CPU Xeon 3.16 GHz used to measure the time of encoding.

The results of experiments on the degree of increase in bit coding and the degree of reduction time encoding is shown in table 1.

Table 1
The degree of reduction time encodingThe degree of increase bit coding
City22,97%0,62%
Football21,63%0,80%
Pedestrian18,80%0,77%
Station21,46%0,11%

Here, the degree of increase bit coding created by performing piecewise cubic Hermite polynomial interpolation for approximating the curve between the four plotted points for a bit of coding and PSNR (peak signal-to-noise ratio) in each QP, and the average differential value of the value code in shared segments of these two compared data sets are specified as relevant degree uvelicheniya, the degree of reduction of encoding time is the average value in each QP.

Characteristics coding for image coding "Pedestrian" and "Station" shown in Fig, and changes in the time of encoding when encoding "Pedestrian" shown in Fig.

From the experimental results above confirmed that, regardless of resolution or image, you can reach the degree of reduction in encoding time of approximately 20%, while the rate of increase of the bit encoding is supported by less than 1%.

The invention is not limited to the above embodiment. For example, in the above embodiment describes an example in which the present invention is applied to an ordered multilayer structure formed through the base layer and improving layer, however, applications of the present invention is not limited to this type of an ordered multilayer structure.

Industrial applicability

The present invention can be applied to scalable coding of moving images, which achieves scalability through the use of a multilayer structure, and can reduce the encoding time by applying the present invention.Reference number

101 module storing initial values var the ants search mode prediction

102 - detection search mode prediction

103 module determine the optimal prediction modes

104 module determine the optimal prediction modes

105 module generate residual signals of forecasting

106 module encoding the residual signal prediction

107 - determining module complete coding MB

108 module updates the target MB coding

201 module storing information specifying the target MB narrowing the search mode prediction

202 module defining the target MB narrowing the search mode prediction

203 module storing information specifying a target frame calculate the degree of match modes forecasting

204 module storing the optimal modes of forecasting improving layer for the target frame

205 module storage for optimal prediction mode for the layer directly under the target frame

206 module create tables of degrees of match modes forecasting

207 module for storing tables of degrees of match modes forecasting

208 module storing threshold values of the narrow search mode prediction

209 module comparing the threshold value table of the degrees of match modes forecasting

301 - module storing information specifying the target MB narrowing the search mode prediction

302 - module defining the target MB narrowing the search mode prediction

303 - module storing information specifying a target frame calculate the degree of match modes forecasting

304 - module storing the optimal modes of forecasting improving layer for the target frame

305 - module storage for optimal prediction mode for the layer directly under the target frame

306 - module create tables of degrees of match modes forecasting

307 - module storing tables of degrees of match modes forecasting

308 - module analysis-prediction mode to the maximum fraction of precipitation

1. How scalable video coding for scalable video coding, comprising:
- the stage at which, on the basis of information about the optimal prediction mode that was selected when scalable coding is performed without imposing restrictions on the use of prediction modes specified as available for use, are shares of correlation combinations of optimum prediction modes between levels, which should be selected for spatially corresponding blocks of the upper layer and the lower layer, and creates a lookup table that describes the relationship between selected combinations of the optimum of the prediction mode and the optimal prediction mode, which should be selected, and the above-mentioned share of correlation, and the lookup table stores the share of conformity for all combinations of modes internal prediction modes mutual prediction and modes of inter-layer prediction is used for the selected optimum prediction mode and the optimal prediction mode that should be selected;
- the stage at which, when the coding block of the upper layer, reveals information about the optimal prediction mode that is selected when the spatial encoding of the corresponding block of the lower layer; and
- the stage at which, on the basis of information about the selected optimum prediction mode detected in the detection step, and on the basis of information about the shares of the correlation described in the correlation table is extracted effective combination from among combinations described in the table, and the optimum prediction mode of the upper layer contained in the extracted effective combination, is defined as the search variant of the prediction mode, the search must be performed when encoding a block of the upper layer.

2. How scalable video coding for scalable video coding, comprising:
- the stage at which, on the basis of information about the optimal mode cons the licensing, which was selected when scalable coding is performed without imposing restrictions on the use of prediction modes specified as available for use, are shares of correlation combinations of optimum prediction modes between levels, which should be selected for spatially corresponding blocks of the upper layer and the lower layer, and creates a lookup table that describes the relationship between the combinations of the selected optimum prediction mode and the optimal prediction mode that should be selected, and the above-mentioned share of correlation, and the lookup table stores the share of conformity for all combinations of modes internal prediction modes mutual prediction and modes of inter-layer prediction is used for the selected optimal mode forecasting and optimal prediction mode that should be selected;
- the stage at which, on the basis of values of the degree of correlation is extracted the effective combination of optimum prediction modes, by narrowing the combinations of the selected optimum prediction modes described in the table, and the optimal prediction mode that should be selected, and generates information of compliance for the forecast of the simulation, which describes the combinations of the extracted effective optimal prediction modes;
- the stage at which, when the coding block of the upper layer, reveals information about the optimal prediction mode that is selected when the spatial encoding of the corresponding block of the lower layer; and
- the stage at which, by reference to the information compliance for forecasting using, as key information about the selected optimum prediction mode detected in the detection step, is determined by the search variant of the prediction mode, the search must be performed when encoding a block of the upper layer.

3. The method of scalable video coding according to claim 2, in which:
- at the stage of creating information compliance for forecasting, the combination, with a share of correlation, showing the value that exceeds a predetermined threshold value is extracted as an effective combination.

4. The method of scalable video coding according to claim 2, in which:
- at the stage of creating information compliance for forecasting, the combination, with a share of correlation, showing the highest value from among combinations of optimum prediction modes that have identical optimal prediction modes to lower the third layer, is extracted as an effective combination, or, alternatively, a combination of a predetermined number of optimal prediction modes selected sequentially based on the proportion of correlation, showing the highest value, is extracted as an effective combination.

5. The device is scalable video coding, scalable way, which encodes the video that contains:
a generation module that, based on the information about the optimal prediction mode that was selected when scalable coding is performed without imposing restrictions on the use of prediction modes specified as possible to use, determines the share of correlation combinations of optimum prediction modes between levels, which should be selected for spatially corresponding blocks of the upper layer and the lower layer, and creates a lookup table that describes the relationship between the combinations of the selected optimum prediction mode and the optimal prediction mode that should be selected, and the above-mentioned share of correlation, and the lookup table stores the share of conformity for all combinations of modes internal forecasting, modes of mutual prediction and modes of interlayer prediction used to select the tion of the optimal prediction mode, and for the optimal prediction mode that should be selected;
- detection module, which, when encoding a block of the upper layer, detects information about the optimal prediction mode that is selected when the spatial encoding of the corresponding block of the lower layer; and
- determining module that, based on the information about the selected optimum prediction mode detected by the detection module, and on the basis of information about the shares of the correlation described in the table of correspondence, extracts the effective combination from among combinations described in the lookup table, and determines the optimum prediction mode of the upper layer contained in the extracted effective combination, as the search variant of the prediction mode, the search must be performed when encoding a block of the upper layer.

6. The device is scalable video coding, scalable way, which encodes the video that contains:
module create lookup tables, which, on the basis of information about the optimal prediction mode that was selected when scalable coding is performed without imposing restrictions on the use of prediction modes specified as possible to use, determines the share of correlation of optimal combinations of modes is prognozirovaniya between levels, who should be selected for spatially corresponding blocks of the upper layer and the lower layer, and creates a lookup table that describes the relationship between the combinations of the selected optimum prediction mode and the optimal prediction mode that should be selected, and the above-mentioned share of correlation, and the lookup table stores the share of conformity for all combinations of modes internal prediction modes mutual prediction and modes of inter-layer prediction is used for the selected optimum prediction mode and the optimal prediction mode that should be selected;
- creation module information compliance for forecasting, which is based on the values of degree of correlation, extracts the effective combination of optimum prediction modes by narrowing combinations of the selected optimum prediction modes described in the table, and the optimal prediction mode that should be selected, and creates information compliance for forecasting, which describes the combination of the extracted effective optimal prediction modes;
- detection module, which, when encoding a block of the upper layer, detects information about the optimal d is the ima forecasting, selected when the spatial encoding of the corresponding block of the lower layer; and
module definitions, which, by reference to the information compliance for forecasting using, as key information about the selected optimum prediction mode detected by the detection module determines the search mode prediction, the search must be performed when encoding a block of the upper layer.

7. The computer-readable recording medium that stores the program scalable video coding, which is used to perform a computer method for scalable video encoding, described in any one of claims 1 to 4.



 

Same patents:

FIELD: information technology.

SUBSTANCE: displacement vectors are searched for by searching for global displacement, breaking up the image into multiple layers of blocks, successive processing of the layers using various search schemes, using displacement vector prediction, as well as selecting displacement vectors based on efficiency of their further entropy coding.

EFFECT: quality improvement of efficiency of a video compressing system, especially at low bit losses, high output thereof.

2 cl, 8 dwg

FIELD: information technologies.

SUBSTANCE: video coding device is a video coding device for exposure of a video image to forecasting coding with compensation of motion, comprising a detection module, in order to detect accessible blocks for blocks having vectors of motion, from coded blocks adjacent to a block to be coded, and a number of available blocks, a selection module, in order to select one selective block from coded accessible blocks, a coder of selection information, to code information of selection, indicating the selective block, using a coding table, corresponding to the number of accessible blocks, and a coder of images, to expose the block to be coded to forecasting coding with compensation of motion using a vector of motion of the selective block.

EFFECT: reduction of additional information by information of selection of a motion vector with increased extents of freedom for calculation of a motion vector by selection of one of coded blocks.

10 cl, 14 dwg

FIELD: information technology.

SUBSTANCE: each re-encoded frame of a multiview video sequence, defined according to a predetermined encoding sequence, is presented as a set of non-overlapping units; at least one of already encoded frame is determined, which corresponds to said view and denoted as reference; synthesised frames are generated for the encoded and reference frames, wherein for each non-overlapping unit of pixels of the encoded frame, denoted as the encoded unit, a spatially superimposed unit inside the synthesised frame is determined, which corresponds to the encoded frame, denoted as a virtual unit, for which the spatial position of the unit of pixels in the synthesised frame which corresponds to the reference frame is determined, so that the reference virtual unit thus determined is the most accurate numerical approximation of the virtual unit; for the determined reference virtual unit, the spatially superimposed unit which belongs to the reference frame, denoted as the reference unit, is determined, and the error between the virtual unit and the reference virtual unit is calculated, as well as the error between the reference virtual unit and the reference unit; the least among them is selected and based thereon, at least one differential encoding mode is determined, which indicates which of the units found at the previous should be used to perform prediction during the next differential encoding of the encoded unit, and differential encoding of the encoded unit is carried out in accordance with the selected differential encoding mode.

EFFECT: providing differential encoding of a frame using a small volume of service information by taking into account known spatial connections between neighbouring views at each moment in time, as well as information available during both encoding and decoding.

5 cl, 15 dwg

FIELD: information technology.

SUBSTANCE: method of encoding an image using intraframe prediction involves selecting a pixel value gradient which is indicated by the image signal to be predicted from among a plurality of selected gradients; generating a predicted signal by applying the gradient in accordance with the distance from the reference prediction pixel, based on the gradient; intraframe encoding of the image signal to be predicted, based on the predicted signal; and encoding information which indicates the value of the selected gradient. As an alternative, the method involves estimating the pixel value gradient which is indicated by the image signal to be predicted, based on the image signal already encoded; generating a predicted signal by applying the gradient in accordance with distance from the reference prediction pixel, based on the gradient; and intraframe encoding of the image signal to be predicted, based on the predicted signal.

EFFECT: improved image compression efficiency.

20 cl, 55 dwg

FIELD: information technology.

SUBSTANCE: method of encoding a video signal comprises steps of: forming a predicted image for the current block; generating a weighted prediction coefficient for scaling the predicted image; forming a weighted prediction image by multiplying the predicted image with the weighted prediction coefficient; generating a difference signal by subtracting the weighted prediction image from the current block; and encoding the difference signal, wherein generation of the weighted prediction coefficient involves calculating the weighted prediction coefficient for which the difference between the base layer image, which corresponds to the current block, and the predicted image is minimal.

EFFECT: high efficiency of encoding a video signal by reducing the error of the current block, which must be compressed, and the predicted image.

31 cl, 16 dwg

FIELD: information technology.

SUBSTANCE: deblocking filter 113 adjusts the value of disable_deblocking_filter-idc, slice_alpha_c0_offset_div2 or slice_beta_offset_div2 based on the Activity of an image calculated by an activity calculation unit 141, the total sum of orthogonal transformation coefficients of the image calculated by an orthogonal transformation unit 142, Complexity of the image calculated by the rate control unit 119, or the total sum of prediction errors of the image calculated by a prediction error addition unit 120.

EFFECT: improved image quality through correct deblocking.

8 cl, 7 dwg

FIELD: information technology.

SUBSTANCE: disclosed is an image decoding method comprising steps of parsing network abstraction layer (NAL) units of a base view (S200); decoding an image of the base view (S202); parsing multiview video coding (MVC) extension parameters of a non-base view (S204); searching whether or not prefix NAL units for a base view are present (S205); either calculating MVC extension parameters for the base view when no prefix NAL units are present (S206) or parsing the MVC extension parameters of the base view when prefix NAL units for the base view are present (S207); and decoding the non-base view using the MVC extension parameters of the base view and the MVC extension parameters of the non-base view (S210).

EFFECT: providing multiview video coding methods of multiview video decoding methods, even when prefix NAL units are not used.

2 cl, 23 dwg

FIELD: information technologies.

SUBSTANCE: method of video coding includes establishment of candidates of reference pixels for pixels within a previously specified range of distances measured from a target coding unit; generation of a predicted signal by means of serial selection of reference pixels used for inner prediction of the target coding unit, among reference pixels-candidates, whenever a condition of distance from the target coding unit varies, and by generation of a predicted signal by reference pixels for each condition of distance; calculation of costs for coding to implement coding with inner prediction of the target coding unit using each generated predicted signal; final detection of reference pixels used for inner prediction of the target coding unit, on the basis of each calculated cost for coding; and coding of information indicating position of detected reference pixels.

EFFECT: provision of efficient internal prediction of an image, which contains eclipses or noise, or to an image, where signals arise that have similar spatial frequencies, such images may not be processed by means of a regular internal prediction.

10 cl 18 dwg

FIELD: information technologies.

SUBSTANCE: image coding device determines a basic factor of quantisation (QPmb), serving as a predictable quantisation factor in case, when a generated code of the main coding approaches the target size of the code during coding of the input image. The image coding device codes the input image for each increment during control with feedback by performance of quantisation with the help of an adapted quantisation parameter (QPt) on the basis of an average parameter.

EFFECT: invention provides for compression of a generated code size for each image increment below the target size of a code in a proper manner without use of a quantisation factor serving as the basis for a quantisation pitch, with high deviation.

14 cl, 7 dwg

FIELD: physics, communication.

SUBSTANCE: invention relates to picture digital signal encoder/decoder in appropriate chroma signal format. Encoder comprises module to divide input bit flow into appropriate colour components, module to divide colour component input signal into blocks so that coding unit area signal can be generated, module to generate P-frame for said signal, module to determine prediction mode used for encoding by efficiency P-frame prediction, module to encode prediction error for encoding difference between predicted frame corresponding to prediction mode as defined by appropriate module and colour component input signal, and module for encoding with variable length of prediction mode code, output signal of prediction error encoding module and colour component identification flag indicating colour component wherein belongs said input bit flow resulted from division of colour components.

EFFECT: higher efficiency.

2 cl, 98 dwg

FIELD: video decoders; measurement engineering; TV communication.

SUBSTANCE: values of motion vectors of blocks are determined which blocks are adjacent with block where the motion vector should be determined. On the base of determined values of motion vectors of adjacent blocks, the range of search of motion vector for specified block is determined. Complexity of evaluation can be reduced significantly without making efficiency of compression lower.

EFFECT: reduced complexity of determination.

7 cl, 2 dwg

FIELD: compensation of movement in video encoding, namely, method for encoding coefficients of interpolation filters used for restoring pixel values of image in video encoders and video decoders with compensated movement.

SUBSTANCE: in video decoder system for encoding a video series, containing a series of video frames, each one of which has a matrix of pixel values, interpolation filter is determined to restore pixel values during decoding. System encodes interpolation filter coefficients differentially relatively to given base filter, to produce a set of difference values. Because coefficients of base filter are known to both encoder and decoder and may be statistically acceptably close to real filters, used in video series, decoder may restore pixel values on basis of a set of difference values.

EFFECT: efficient encoding of values of coefficients of adaptive interpolation filters and ensured resistance to errors of bit stream of encoded data.

5 cl, 17 dwg

FIELD: video encoding, in particular, methods and devices for ensuring improved encoding and/or prediction methods related to various types of video data.

SUBSTANCE: the method is claimed for usage during encoding of video data in video encoder, containing realization of solution for predicting space/time movement vector for at least one direct mode macro-block in B-image, and signaling of information of space/time movement vector prediction solution for at least one direct mode macro-block in the header, which includes header information for a set of macro-blocks in B-image, where signaling of aforementioned information of space/time movement vector prediction solution in the header transfers a space/time movement vector prediction solution into video decoder for at least one direct mode macro-block in B-image.

EFFECT: creation of improved encoding method, which is capable of supporting newest models and usage modes of bi-directional predictable (B) images in a series of video data with usage of spatial prediction or time distance.

2 cl, 17 dwg

FIELD: movement estimation, in particular, estimation of movement on block basis in video image compression application.

SUBSTANCE: method and device are claimed for conducting search for movement in video encoder system using movement vectors which represent difference between coordinates of macro-block of data in current frame of video data and coordinates of corresponding macro-block of data in standard frame of video data. A set of movement vector prediction parameters is received, where movement vector prediction parameters represent approximations of possible movement vectors for current macro-block, movement vector search pattern is determined and search is conducted around each movement vector prediction parameter from the set of movement vector prediction parameters using search pattern, and on basis of search result, the final movement vector is determined.

EFFECT: increased efficiency of video signals compression.

3 cl, 7 dwg

FIELD: physics.

SUBSTANCE: said utility invention relates to video encoders and, in particular, to the use of adaptive weighing of reference images in video encoders. A video encoder and a method of video signal data processing for an image block and the specific reference image index for predicting this image block are proposed, which use the adaptive weighing of reference images to increase the video signal compression, the encoder having a reference image weighting factor assigning module for the assignment of the weighting factor corresponding to the said specific reference image index.

EFFECT: increased efficiency of reference image predicting.

8 cl, 7 dwg

FIELD: physics, computing.

SUBSTANCE: invention relates to the field of coding and decoding of a moving image. In the method, at least one reference image for the processing of the field macroblock is selected from at least one reference image list, using information about reference image indexes, each at least one reference image selected is a field, and the parity of at least one reference field selected may be based on the parity of the field macroblock and the reference image index information.

EFFECT: efficient provision of information about reference image compensating motion, by reference image indexes determined in different ways, according to the coded macroblock modes.

10 cl, 12 dwg

FIELD: information systems.

SUBSTANCE: invention refers to video coders using adaptive weighing of master images. The video decoder for decoding data from video signal for the image, having multiple motion boxes, containing: the master image weighting coefficient module for accepting, at least, one master image index, thereat each one from the mentioned master image indexes is intended for independent indication not using any other indexes, one of the multiple master images, used for prediction of current motion box and weighting coefficient from the set of weighting coefficients for current mentioned one from mentioned multiple motion boxes.

EFFECT: increase of efficiency in predicting master images.

20 cl, 7 dwg

FIELD: information technology.

SUBSTANCE: method is offered to compress digital motion pictures or videosignals on the basis of superabundant basic transformation using modified algorithm of balance search. The algorithm of residual energy segmentation is used to receive an original assessment of high energy areas shape and location in the residual image. The algorithm of gradual removal is used to decrease the number of balance assessments during the process of balance search. The algorithm of residual energy segmentation and algorithm of gradual removal increase encoding speed to find a balanced basis from the previously defined dictionary of the superabundant basis. The three parameters of the balanced combination form an image element, which is defined by the dictionary index and the status of the basis selected, as well as scalar product of selected basic combination and the residual signal.

EFFECT: creation of simple, yet effective method and device to perform frame-accurate encoding of residual movement on the basis of superabundant basic transformation for video compressing.

10 cl, 15 dwg

FIELD: information technology.

SUBSTANCE: playback with variable speed is performed without picture quality deterioration. Controller 425 creates EP_map () with RAPI address in videoclip information file, dedicated information selection module 423 RAPI, image PTS with internal encoding, which is immediately preceded by RAPI, one of final positions of the picture with internal encoding, as well as the second, the third and the fourth reference pictures, which are preceded by the picture with internal encoding. The controller saves EP_map () in output server 426, i.e. controller 425 copies the value, close to given number of sectors (quantity of sectors, which can be read at one time during encoding process) of final positions for the four reference pictures (1stRef_picture, 2ndRef_picture, 3rdRef_picture and 4thRef_picture) to N-th_Ref_picture_copy, defines value of index_minus 1 on the basis of N-th_Ref_picture_copy and records it to disc.

EFFECT: effective process performance with constant data reading time.

8 cl, 68 dwg

FIELD: information technology.

SUBSTANCE: invention proposed contains videodecoder (200) and corresponding methods of videosignal data processing for image block with two reference frames' indices to predict this image block. The methods use latent scaling of reference images to improve video compressing. The decoder (200) contains latent scaling coefficient module (280) of reference images, which are used to determine a scaling coefficient value, corresponding to each of the reference image indices. Decoding operations contain receiving reference image indices with data, which corresponds to image block, calculation of latent scaling coefficient in response to image block location relative to reference images, indicated by each index of reference image, extraction of reference image for each of the indices, motion compensation relative to extracted reference image and multiplication of reference images, relative to which the motion compensation was performed, to a corresponding scaling value.

EFFECT: increase of decoding efficiency.

25 cl, 6 dwg

Up!