Device for coding of images, device for decoding of images, method for coding of images and method for decoding of images

FIELD: information technologies.

SUBSTANCE: decoding stage is carried out to decode the first identifying information, which indicates whether signals of appropriate colour components are independently coded, and to decode the second identifying information in the case when the first identifying information indicates that signals of appropriate colour components are coded independently, the second identifying information indicates whether all colour components are intra-coded, besides, in the case when the first identifying information indicates that signals of the appropriate colour components are coded independently, and the second identifying information indicates that all images of appropriate colour components are intra-coded, the decoding stages generates a decoded image, in which performance of deblocking filtration is off at each border between blocks serving as elements of conversion and quantisation.

EFFECT: improved optimality and efficiency of decoding in case when coded signals of a dynamic image do not have a difference in respect of a number of counts between colour components.

15 dwg

 

The technical field to which the invention relates.

The present invention relates to a device for coding digital image signals, the device for decoding digital signals of images, the method of coding digital image signals and the method of decoding digital signals of images used for encryption and compression of images or data transmission technology compressed images.

The level of technology

The system of coding a video signal of international standard, such as MPEG or ITU-T NH (for example, the standard "Information Technology Coding of Audio-Visual Objects Part 10: Advanced Video Coding", ISO/IEC 14496-10, 2003 (hereinafter referred to as non-patent document 1)), has traditionally been based on the use of a standardized input format called 4:2:0. Format 4:2:0 is the format in which the color signal of the dynamic image in RGB components or similar form is converted into a brightness component (Y) and two components of chrominance (b, SG), and the number of times the component of the color is reduced to half the quantity of the luminance components in both the horizontal and vertical directions. Component color visually varies less than the brightness. Under this traditional system of video coding between the national standard was based on the premise the amount of initial information, which must be coded is reduced by domain downsampling chrominance components before performing the coding, as mentioned above. On the other hand, the recent increase in resolution and number of tones of the video was accompanied by a research system for performing coding by supporting the number of samples equal to the number of the luminance components, without domain downsampling chrominance components. The format in which the number of samples of the luminance components and chrominance fully equal, is called a format of 4:4:4. The traditional format of 4:2:0 was limited by the definitions of the components Y, b and CR color space due to the background domain downsampling chrominance components. However, in case the format is 4:4:4, because there is no difference in the ratio of the number of samples between the color components can be directly used components R, G, and In addition to the components Y, b and SG and can be used many definitions of the color space. An example of encoding a video aimed at format 4:4:4, is to publish Woo-Shik Kim, Dae-Sung Cho and Hyun Mun Kim, "INTER-PLANE PREDICTION FOR RGB VIDEO CODING", ICIP 2004, October 2004 (hereafter referred to as non-patent document 2). Non-patent document 2 proposes an approach to reduce the size of data that D. what should be encoded by prediction between different color components using correlation remaining between the color components. However, the degree of correlation between the color components varies depending on the type of content of the video signal and color space, and the prediction can produce an adverse effect in the context of coding efficiency. In addition, the required signal processing on the multiple color components, thus, there is a problem that affects the efficiency of parallel processing, for example, when performing real-time processing of video signals with extremely high resolution as the digital image of the film (with the resolution of 40002000 pixels).

The INVENTION

In the format of 4:2:0 advanced encoding video signals (hereinafter referred to as encoding AVC) MPEG-4 of non-patent document 1 in the field macroblock composed of the luminance components with a size of 1616 pixels corresponding to the color component are blocks of size 8x8 pixels for both components b and SG. When the prediction with motion compensation format 4:2:0 are multiplexed with information about the block size, which becomes an element of prediction with motion compensation only for the luminance components, inform the tion of the reference image, used for predictions, and information of a motion vector of each block, and the prediction with motion compensation is performed for the chrominance components using the same information as for the luminance components. Format 4:2:0 has such characteristics in the definition of the color space that almost all the items of information about the structure (information about the texture image is integrated into the brightness component of the color visibility distortion lower than for the brightness component, and the contribution to the reproducibility of the signal is small, and the prediction and encoding in the format of 4:2:0 is based on such characteristics. However, in case the format is 4:4:4 three color component have the same texture information. System to perform prediction with motion compensation in the prediction mode with internatioanal, depends only on one component and based on information of the reference image and information of the motion vector is not necessarily the best way in the format of 4:4:4, where the color components make equal contributions to the representation of the structure of the image signal.

As described above in connection with prior art, the present invention is to provide a device for encoding, for decoding, encoding and str is both decoding, which increase the optimality in the case where the encoded dynamic image signals do not have differences in the ratio of the number of samples between the color components, as in the case of format 4:4:4.

In accordance with the present invention in the case of encoding with compression by entering the digital signals of the dynamic image in the format of 4:4:4 prepared by first encoding process of the three signals of the color components of the input signals of the dynamic image mode General encoding and the second encoding process of the three signals of the color components of the input dynamic image signals in the respective modes are independent of the encoding. The encoding process is performed by selecting any process from the first encoding and the second encoding process, and data compression include an identifying signal to indicate which process is selected.

In addition, in the case of decoding data compression of the digital signals of the dynamic image in the format of 4:4:4 prepared by first decoding process of the three signals of the color components in the encoding and the second decoding process of the three signals of the color components in the respective modes are independent of the encoding. The decoding process is performed by decoding : CTCSS / DCS the project identifying signal, specifies the coded whether the three color signal components of data compression in encoding or corresponding modes independent coding, and the use of any process from the first process of decoding and the second decoding process in accordance with the identifying signal.

In accordance with the present invention in the case of performing encoding that uses multiple color spaces, not limited to fixed color spaces with components Y, b, SG, etc., it is possible to flexibly select the information mode prediction with internatioanal that will be used for the respective color components, and to perform the optimum encoding process even in the case where there are a variety of definitions of color spaces.

BRIEF DESCRIPTION of DRAWINGS

In the accompanying drawings:

1 is an explanatory diagram showing a hierarchical structure of a video signal consisting of a sequence, group of pictures (GOP), picture, sections and macroblock;

2 is an explanatory diagram showing the structure of the encoded data of the macroblock;

3 is an explanatory diagram showing three color component forming the structure of the macroblock in the case of "process complete coding";

4 is an explanatory CX is mA showing three color component forming the structure of the macroblock in the case of the process of independent coding";

5 is an explanatory diagram showing a reference relationship for the directed time-motion prediction between pictures in the "common coding" and "independent encoding";

6 is an explanatory diagram showing an illustrative structure of a bit stream in accordance with a first variant embodiment of the present invention;

7 is an explanatory diagram showing an illustrative structure of the bitstream data section in accordance with the first variant embodiment of the present invention;

Fig is a schematic structural diagram of a device for coding according to the first variant embodiment of the present invention;

Figure 9 - diagram of the internal configuration of the first unit 5, the image encoding;

Figure 10(a)-10(h) is an explanatory diagram showing the seven types of dimensions, each of which is the prediction with motion compensation;

11 is an internal structural diagram of the second unit 7, the image encoding;

Fig is a schematic structural diagram of a device for decoding in accordance with the first variant embodiment of the present invention;

Fig - internal structural diagram of the first BL is ka 302 decoded image;

Fig - internal structural diagram of the second unit 304 decodes the image; and

Fig is an explanatory diagram showing an example of how the distributed image data for the three color components.

The IMPLEMENTATION of the INVENTION

The first variant embodiment of the invention

The first variant of the embodiment will illustrate a device for encoding, which effectively encodes the video signal in the format of 4:4:4, is not limited to a specified color space, and a device for decoding, which receives the encoded bit streams generated by the device for coding and decoding image signals. Device for encoding in the first variant embodiment receives a video signal containing three color component, such as RGB, XYZ and YCbCr, and generates bit streams, performing encoding with compression. As illustrated in figure 1, the input signals are represented as time-series data on-screen information (hereinafter called image), defined for each frame or for each field through the discretization in time. The data item on which the images are ordered in the time series, is called a sequence. The sequence can be divided into some groups of pictures (GOP). Group of pictures (GOP) are used for such a the th, as enforcement decoding, starting from arbitrary initial group of pictures (GOP), independently of other groups of pictures (GOP) and providing random access to the bit stream. The image is divided into square blocks, called macroblocks, and the compression of the video signal is performed at the macroblock using the prediction process/transformation/quantization.

In addition, the element in which United many of the macroblocks is called a section. A partition is a data item, which can be performed independently of the encoding and decoding in each section. For example, when processing video signals with a resolution equal to or greater resolution high-definition television (HDTV), real-time one image is cut into many sections, the period calculation time is reduced through parallel encoding/decoding of the respective sections, and the bit streams are transmitted through a line having a high error rate. In this case, the section is used with the intention that, if the decoded image is distorted due to the destruction of the section affected by the error, correct the decoding process is restored with the next section. In the General case, the prediction using the dependence of the signal on neighboring the section cannot be applied at the boundary between the sections, and therefore the encoding performance decreases with increasing number of partitions. However, there are such characteristics as flexibility of parallel processing, and the ability to recover from errors increases.

The compression of the video signal is performed by applying the process of prediction/transform/quantization at the level of the macroblock, and therefore, the encoded data of the macroblock, multiplexed in the bit stream are, roughly speaking, of two types of information. One type of information is a category, called auxiliary information that is different from the video signal itself, for example, parameters for the prediction mode, information of the motion prediction and quantization, and this category of information in the General case is called the header of the macroblock. Another type of information is information of the video signal. In accordance with the first variant embodiment of the present invention, the video signal which is to be encoded, is a compressed data signal with the error of the predictions obtained in the result of the prediction/transform/quantization on the basis of header information of the macroblock. The video is presented in the format of the quantization conversion efficiency, and therefore will continue to be called on nami, encoded coefficient. Figure 2 illustrates how there are elements of the encoded data of the macroblock in the first variant embodiment of the invention. Figure 2 macroblock header contains all of the elements supporting information prior to the data encoded by the coefficient, such as information encoding mode/prediction of the macroblock type/type submicromolar/mode prediction with intracoronal etc., information of the motion prediction about the number identifying the reference image/displacement vector, etc., the quantization parameter relative to the conversion factor and a flag indicating the presence/absence of the applicable conversion factor for each block of size 8x8.

The first variant of the embodiment of the present invention will illustrate: a device for encoding, which encodes three color component selectively on the basis of common header macroblock-based or independent header macroblock in accordance with each color component; and a device for decoding, which performs the decoding process of the video signal by receiving bit streams received at the output device for encoding, and determining on the basis of the identifying information extracted from the bitstream by decoding foogo thread there were three color component is encoded on the basis of common header macroblock-based or independent header macroblock for each color component. The first variant of the embodiment of the present invention, in particular, illustrate with reference to some drawings of the configuration and the operation of the device for encoding and device for decoding that perform encoding and decoding, switching on the sequence level between the operation of encoding and decoding signals from the three color components on the basis of common header of the macroblock and the operation of encoding and decoding these color signals on the basis of independent header macroblock corresponding to each color component. Thus, a device for encoding and device for decoding to encode the color components by selecting either case the encoding of the color components in accordance with the color space in which the input video signals, and in accordance with the statistical characteristics of the video signals using the General parameters of the prediction, or the case of encoding the color components using independent parameters prediction. Therefore, it is possible to optimally encode the video in the format of 4:4:4.

Further about the ECC encoding of the signals of the three color components of one frame or one field through common header macroblock is called "process complete coding", and the process of coding the three color components of one frame or one field by an individual independent of the macroblock header is called the "process of independent coding". Similarly, the decoding process of the image data of the bit streams, in which the signals of the three color components of one frame or one field is encoded through a common header macroblock, called a "General process of decoding, and the decoding process of the image data of the bit streams, in which the signals of the three color components of one frame or one field coded by an individual independent of the header macroblock, called a "process independent decoding". In the General process of encoding three color component of the one frame or one field all together defined as one image, and is divided into macroblocks, each of which consists of combining the three color components (figure 3). Figure 3 and in the following description of these three color component are referred to as components C0, C1 and C2. On the other hand, in the process of independent coding the input video signal of one frame or one field is divided into three color components, each of which is defined as an image, and each image is divided into macroblocks, comprising one svetaveretenenko (figure 4). In other words, the macroblock as a General purpose encoding contains the samples (pixels) of the three color components C0, C1 and C2, but macroblock as the goal of the process is independent of the encoding contains only the samples (pixels) of any one component C0, C1 and C2.

Figure 5 illustrates the relation of reference for motion prediction in the time direction between the images in the device for encoding/device for decoding in accordance with the first variant embodiment of the present invention. In this example, the data item indicated by the thick vertical line represents the image, and the relation between the image and the element of access indicated by the dotted line. In the case of the General process of encoding/decoding, as described above, one image is data representing a video signal, consisting of associations of three color components. In the case of the independent encoding/decoding one picture is a video consisting of any one color component. Element access is a minimum data element for providing a time-stamp, designed, for example, to synchronize video with audio information, etc. In the case of the General process of encoding/decoding a single element access contains the data is for a single image. On the other hand, in the case of the independent encoding/decoding one element access contains three images. The reason for this is that in the case of the independent encoding/decoding video update for one frame will not be received until the images are available for all three color components for the same display time. You should note that the number on top of the corresponding image point aimed in time sequence encoding/decoding (which corresponds to the frame_num coding AVC) images. Figure 5 arrow between images indicates the direction of reference for motion prediction. More specifically, in the case of the independent encoding/decoding reference for motion prediction between images contained in the same element access reference for motion prediction between different color components are not executed, but the process of encoding/decoding are performed when executing the references for predicting limit images to color components C0, C1 and C2 signals of the same one color component. With this structure, in the case of the independent encoding/decoding in accordance with the first variant embodiment of the present izobreteny what each color component can be encoded and decoded independently from the process of encoding/decoding of other color components, thereby parallel processing.

The identification information indicating whether the coding is based on the General process of coding or coding on the basis of independent encoding process is hereinafter referred to as signal 1 identify common encoding/independent encoding.

6 shows one example of the structure of the bitstream generated by the device for coding according to the first variant embodiment of the present invention and serving as a target for the process of receiving/decoding through the device for decoding in accordance with the first variant embodiment of the invention. 6 illustrates the structure of the bit stream from the sequence level to the level of the section. First of all, the title of high order on the sequence level (parameter set sequence in the case of encoding AVC) multiplexed signal 1 identify common encoding/independent encoding. The separator AUD is an item level of network abstraction layer (NAL) for separator access, which is a unique element of the NAL level to identify separator access when encoding AVC. The separator AUD is information that defines the start of the element access, and regardless of the format is given the s separator AUD AVC encoding can be applied to arbitrary data format, if it meets its purpose. For example, the separator AUD corresponds to the initial code image in accordance with the standards MPEG-2 and the initial code of the video object plane (VOP) in accordance with the standards of MPEG-4.

When the signal 1 identify common encoding/independent encoding indicates the process is complete coding", element access contains the encoded data for one image. The image at this time is data representing the video signals for one frame or one field, which consist of associations of three color components, as described above. The actual data encoded video signals multiplexed in the bit stream for each section, shown in figure 1. On the other hand, when the signal 1 identify common encoding/independent encoding specifies the process of independent coding", one image is a video for any one color component in one frame or one field, and one element of access contains three images. In this case, the section is defined relative to the image of each color component.

7 illustrates a structure of bit streams of data sections in the case of the General process of coding and in the case of the process of independent coding, respectively. In the bit stream encoded with the aid of the completion process of independent coding, in order to achieve the effect, which will be described later, for the field header at the beginning of the data section is flag 2 identification color component (color_channel_idc), so that section was distinguishable on which the image of the color component data belong sections taken by the device for decoding. Sections that have the same value are grouped with 2 flag identification. In other words, between sections with different values of flag 2 identification color component, do not enter no dependence encoding/decoding (for example, the reference for motion prediction, the study of probability modeling/build the context for encoding SAVAS and so on). This rule guarantees the independence of the individual images within the element access in the case of the independent encoding. In addition, the frame number (the processing sequence encoding/decoding image belongs section), multiplexed in each header section, takes the same value in the full-color image components within a single element access.

Fig illustrates a schematic configuration of a device for coding according to the first variant embodiment of the present invention. On Fig the first block 5 encoding of the image is agenia performs a process common coding and the second blocks 7a-7C, the image encoding (provided for the three color components) through the process of independent coding.

Video inputs 3 served on any member of a set consisting of a first unit 5, the image encoding unit 6 separation of the color components and the second blocks 7a-7C, the image encoding via the switch 4. Switch 4 is activated by a signal 1 identify common encoding/independent encoding and supplies the input video signal 3 on the specified path.

Signal 1 identify common encoding/independent encoding is a signal that, when the input signal takes the format of 4:4:4, multiplexed in a set sequence and selects the General process of coding and the process of independent coding for each sequence. Signal 1 identify common encoding/independent encoding multiplexed in the parameter set sequence in the bit stream 11 as information that determines what process was used to generate the bit stream 11. This structure allows the device to decode, which takes the bit stream 11, to perform the General process of decoding, when the bit stream 11 is formed by a process common Kodirov the tion, by decoding signal 1 identify common encoding/independent encoding parameter set sequence and check its value, and to perform a process of independent decoding, when the bit stream 11 is formed by a process independent of the encoding.

If the signal is 1 identify common encoding/independent encoding indicates the process is complete coding", in the first block 5, the image encoding an input video signal 3 is divided into macroblocks in the form of combining samples of the three color components, as illustrated in figure 3, performs the encoding process for each macroblock and the encoded data is output as a bit stream 8. The encoding process by the first unit 5, the image encoding will be described later.

If the signal is 1 identify common encoding/independent encoding refers to the process of independent coding", the input signals 3 are divided into signals of color components C0, C1 and C2 through the block 6 separation of color components and is served on the second blocks 7a-7C, the image encoding corresponding to the color components. Secondly blocks 7a-7C, the image encoding signals separated according to each color component is divided into macroblocks, the host format, showing the config in figure 4, and performs the encoding process for each macroblock, whereby the signals are output as bit streams 9a-9c. The encoding process by the second blocks 7a-7C, the image encoding will be described later.

In block 10 of the multiplex signal 1 identify common encoding/independent encoding is added to the set of sequence parameters and is multiplexed in the bit stream 11. In block 10 MUX selects any stream from a set consisting of a bit stream of 8 bit and threads 9a-9c, corresponding to the signal value 1 identify common encoding/independent encoding, and the selected bit stream is multiplexed in the bit stream 11.

In addition, although details will be described later, information (12A-12C) of the weight coefficients quantization to the quantization parameters used in the encoding process of the image, especially in the process of quantization/inverse quantization, is provided for each of the three color components and is input to the respective blocks of the image encoding, which is the quantization process corresponding to the characteristic of each color component. Weights 12A-12C quantization also send the block 10 multiplexing for multiplexing the set of pairs is m sequence, the device for decoding used the same values of the coefficients 12A-12C, which have been used in the encoding process.

In addition, the input unit 5 and 7a-7C, the image encoding signal 13 specifying only intracoronary, thereby controlling the encoding process. Signal 13 specifying only intracoronary is a signal which determines whether the block of the image encoding process of directed time prediction based on prediction with motion compensation. If the signal 13 specifying only intracoronary indicates only interaktywna", is enclosed within the image coding on all images video inputs 3 without performing directional time prediction based on prediction with motion compensation. In addition, in this case, simultaneously disconnects deblocare filter cycle intracoronary (details will be described later) in the block coding of the image. If the signal 13 specifying only intracoronary indicates "not only interaktywna", the process intermediaware using correlation between intracoronary and internatioanal images is relative to the images of input video signals 3 through the use also directed the on-time prediction based on prediction with motion compensation. Unit 10 multiplexing adds the signal 13 specifying only intracoronary to set the sequence settings and, thus, multiplexes the parameter set sequence with the bit stream 11. In this multiplexing device for decoding, the receiving bit stream 11, decodes the signal 13 specifying only intracoronary contained in the parameter set sequence, and checks its value, whereby it is possible to recognize whether the bit stream 11 is encoded using only intracoronary or not. Therefore, if the bit stream 11 is encoded using only intracoronary, execution deblokiruyuschee filtration cycle intracoronary may be unnecessary, and the number of calculations in the device for decoding can be reduced.

The process of intracoronary to encode AVC requires approximately 2-10 times more computation coding in comparison with the process of intermediaware, and therefore data encoded by only intracoronary have considerably larger than the data encoded by "not only intracoronary".

A traditional device for decoding provides an upper limit on the size of the data, allowing for the execution of the decoding process, eumenidae speed and the required size of the memory device in the most possible extent, thereby stabilizing the operation. Therefore, in the case of "only intracoronary" there is a possibility that can be entered which exceeds the upper limit, resulting in a problem in which it is impossible to determine if you can run stable job or not.

With that in mind, the set of parameters sequence includes a flag to indicate whether the size of the encoded data is less than the predetermined number or greater. On the basis of the flag is the process of determining, and if the size of the encoded data is less than the predetermined number, it is assumed that even traditional device for decoding can handle, and so is the process of decoding. If the size of the encoded data exceeds a predetermined amount, it is assumed that the conventional device for decoding may not perform steady process, and therefore can be performed such process, as issuing warnings.

In addition, information 14 about the size of the input image signal is input to blocks 5 and 7a-7C, the image encoding and configured to control the encoding process. Information 14 about the size of the image is information representing the number of macroblocks with intracoronary image is eniami input video signal 3, managed to set the upper limit to the number of macroblocks contained in the section, if the value of the information 14 more predetermined threshold values, and to prevent the section contained more macroblocks than the value of the upper limit. Information 14 about the size of the image is added to the set of sequence parameters, and the parameter set sequence is multiplexed in the bit stream 11. In this multiplexing, if the image size of the input video signal 3 is large (i.e. the spatial resolution is high), and a device for coding and decoding can identify an element that can be processed in parallel, and can uniformly assign tasks.

Next will be explained in detail the work of the first and second units of the image encoding.

The scheme of operation of the first unit of the image encoding

Fig.9 illustrates the internal configuration of the first unit 5, the image encoding. Figure 9 assumes that the input video signal 3 takes the format of 4:4:4 encoded for each macroblock in the format shown in figure 3, in which the macroblock consists of combining the three color components. Internal processing differs depending on the signal values 13 reference is only intracoronary.

(1) the Case in which the signal 13 specifying only intracoronary indicates "not only interaktywna"

The prediction unit 15 selects the reference image data from the reference image for prediction with motion compensation, stored in the memory 16, and performs the process of prediction with motion compensation on the level of macroblocks. In the memory 16 stores the multiple data elements of the reference image, each of which consists of three color components, on the set of the most recent points in time, or past and future moments of time. The prediction unit 15 performs the prediction of the movement, choosing the optimal reference image for each macroblock of these reference images. With regard to the distribution data of the reference image in the memory 16, the data may be stored separately in series planes for each color component, and the counts of the respective color components can also be saved in series points. There are seven types of sizes of blocks that perform prediction with motion compensation. First of all, as illustrated in figure 10(a)-10(d), choose any of the sizes 1616, 168, 816 and 88 for each macroblock. Then, when the selected size 8x8, as illustrated in figure 10(e)-10 (h), for each block of size 88 is selected uboy of sizes 88, 84, 48 and 44. In the General process of encoding performed by the first unit 5, the image encoding for the three color components is chosen and applied, the total block size prediction with motion compensation.

The prediction unit 15 performs the process of prediction with motion compensation for each macroblock with respect to all or part of the dimensions/sizes of the sub-blocks shown in figure 10, the motion vector in a predetermined search range and one or more elements suitable for use reference images, thus giving service information 17 predictions containing the motion vector and the index of the reference picture used for prediction, and the predicted image 33. MyCitadel 18 of the predicted image 33 and the input video signal 3 receives for each block in the differential signal 19 predictions, which serves as a single element prediction with motion compensation. Unit 20 determine the encoding mode performs a selection process based on the prediction process performed by the prediction unit 15, and outputs the selected differential signal 19 predictions and the macroblock type 21/type submicromolar. All the elements of the header information of the macroblock, such as the type of macroblock type submicromolar, the index of the reference image and the vector is viginia, defined as header information common to the three color components. The header information of the macroblock is used to encode the three color components and is multiplexed in the bit stream 8. In the case of an optimality criterion of the effectiveness of forecasting in order to limit the number of calculations can be estimated only the magnitude of the prediction errors for the predetermined color component (for example, a component G of the components R, G and b and the component Y component Y, b and CR), and the magnitude of the prediction errors for all the color components can also be comprehensively evaluated to obtain optimum performance predictions, although the number of calculations increases.

Similarly, the prediction unit 15 performs prediction with intracoronal. When the prediction unit 15 performs prediction with intracoronal, information about a prediction mode with intracoronal displayed on the signal 17.

Further, if there is no specific differences between the prediction with intracoronal and prediction with motion compensation, the signal 17 is called service information predictions. It should be noted that data from the reference image for prediction with intracoronal involve the use of local decode is consistent image 23 before processing debateroom filter 22 (although it is not illustrated, the local decoded image 23 is used as data of a reference image for prediction with intracoronal, and therefore temporarily stored in block 15 predictions and so on). Mode prediction with intracoronal common to the three color components is selected and applied to the General process of encoding performed by the first unit 5, the image encoding. As for predictions with intracoronal, can be estimated the value of the prediction error only for the predetermined color component and can also be comprehensively assessed values of prediction errors for all color components. Finally, the unit 20 determine the encoding mode selects through evaluation from the point of view of efficiency predictions or coding efficiency, whether as a type of the macroblock prediction with intracoronal or prediction with internatioanal.

Block 24 conversion converts the differential signal 19 predictions and outputs the converted signal as the conversion factor for the quantization block 25. In this case, the block size used by a single element to perform a conversion can be selected from many possible sizes such as 4x4 and 8x8. When the size of the transform block is selected, select the config block size when encoding is reflected in the value of the flag 26 destination block size conversion and the flag 26 multiplexed in the bit stream 8. Block 25 quantization produces quantization received at the input of the conversion coefficient on the basis of parameter 27 quantization and weighting coefficients 12A-12C quantization and outputs the quantized result as a quantized coefficient 28 conversion unit 29 coding with variable length and the block 30 of the inverse quantization.

Next will be described the processing unit 25 of quantization. The conversion factor is converted into a signal of a frequency domain signal from the spatial domain by block 24 conversion, is divided into the low frequency region, where the distortion can be clearly seen by the human eye, and the high-frequency region, where the distortion is difficult to detect. Thus, each frequency domain is suspended. In the low frequency range is fine quantization, while in the high frequency region is coarse quantization, whereby it may be made the quantization process, adapted to the human visual perception. Weights 12A-12C quantization are weight parameters set in each frequency region. To convert a block of size 4x4 used 16 weight parameters, and to convert a block size of 88 are used 64 weight parameter As described above, weights 12A-12C quantization multiplexed in a set sequence. However, in the case of implementation of the General process of coding for the three color components use the same weighting coefficient quantization. In line with this, there is no need for multiplexing three factor 12A, 12b and 12C, and enough for multiplexing only one factor. Unit 25 performs quantization the quantization process, in which the transform coefficients of the three color components are weighted using weights 12A-12C quantization, and thus receives the quantized coefficient 28 conversion.

Quantized coefficient 28 conversion for the three color components are encoded using statistical code through unit 29 coding with variable length, using tools such as coding, Huffman and arithmetic coding.

Next, the quantized coefficient 28 conversion is restored in the local decoded differential signal 32 predictions through the block 30 of the inverse quantization and block 31 inverse transformation. Then, in the adder 34, the signal 32 is added to the image 33 of the prediction based on the selected type 21 macroblock/type submicromolar and proprietary information 17 prognose the Finance, and, thus, a local decoded image 23. The local decoded image 23 after processing debateroom filter 22 is stored in memory 16 to be used for further process prediction with motion compensation.

Quantized coefficient 28 conversion, type 21 macroblock/type submicromolar, service information 17 predictions and parameter 27 quantization, which are input to the block 29 coding with variable length, are placed and executed in accordance with a predefined syntax of the encoded data of the macroblock and are packaged (AVC encoding is also referred to as packetization period of the NAL level) in the data entry section, consisting of one macroblock or merging of multiple macroblocks in the format shown in figure 3. Then the data packets are issued as a bit stream of 8.

(2) the Case in which the signal 13 specifying only intracoronary indicates only interaktywna"

The prediction unit 15 performs the steps outlined in Chapter (1) the process of predicting only interactionalism. When the prediction with intracoronal, information of the prediction mode with intracoronal is issued service information 17 predictions. It should be noted that data from the reference image for prediction with intrabody what Finance include using the local decoded image 23 before processing debateroom filter 22 (although it is not illustrated, the local decoded image 23 is used as data of a reference image for prediction with intracoronal and therefore temporarily stored in block 15 predictions). Mode prediction with intracoronal common to the three color components is selected and applied to the General process of encoding performed by the first unit 5, the image encoding. Unit 20 determine the encoding mode selects the mode prediction with intracoronal through evaluation from the point of view of efficiency predictions or coding efficiency.

Unit 20 determine the encoding mode outputs the selected differential signal 19 prediction on a block 24 conversion. Block 24 conversion converts the input differential signal 19 predictions and outputs the converted signal as the conversion factor for the quantization block 25. In this case, the block size used by a single element to perform a conversion can be selected from many possible sizes such as 4x4 and 8x8. The AVC encoding configured in such a way that the target block prediction mode prediction with intracoronal is adjusted according to the size of the transform block. In the case where the block size conversion is done, select, selected when coding the block size of the otra is moved in the value of the flag 26 destination block size conversion and the flag 26 is multiplexed in the bit stream 8. Block 25 quantization produces quantization received at the input of the conversion coefficient on the basis of parameter 27 quantization and weighting coefficients 12A-12C quantization and outputs the quantized result as a quantized coefficient 28 conversion unit 29 coding with variable length. The example process described above.

Quantized coefficient 28 conversion for the three color components are encoded using statistical code through unit 29 coding with variable length, using tools such as coding, Huffman and arithmetic coding.

Next, the quantized coefficient 28 conversion is restored in the local decoded differential signal 32 predictions through the block 30 of the inverse quantization and block 31 inverse transformation. Then, in the adder 34, the signal 32 is added to the image 33 of the predictions based on the service information 17 forecasting, and, thus, a local decoded image 23. If information 13 reference only intracoronary indicates only interaktywna", the prediction with motion compensation is not performed, and therefore deblocare filter 22 does not perform any processing or recording data as a reference from the expressions in the memory 16. With this structure, a memory access and arithmetic operations required to process debateroom filter 22 may be reduced.

Quantized coefficient 28 conversion, type 21 macroblock/type submicromolar 21 (set mode intructional), service information 17 predictions and parameter 27 quantization, which are input to the block 29 coding with variable length, are placed and executed in accordance with a predefined syntax of the encoded data of the macroblock and are packaged (AVC encoding is also referred to as packetization period of the NAL level) in the data entry section, consisting of one macroblock or merging of multiple macroblocks in the format shown in figure 3. Then the data packets are issued as a bit stream of 8.

It should be noted that the number of macroblocks included in the section isn't limited by the value of the information 13 reference only intracoronary and information 14 about the size of the image. Information 14 about the size of the image is used as input information for the block 29 coding with variable length. Block 29 coding with variable length sets on the basis of information 14 about the size of the image is the upper limit of the number of macroblocks included in the section. Block 29 coding with variable length pre-calculates the number of the STV-coded macroblocks, and when the number of macroblocks included in the section reaches the upper limit value, closes the data package section. Subsequent macroblocks pekititirsia how the data in the new partition.

In addition, the first unit 5, the image encoding in accordance with a first variant embodiment of the present invention does not throw the flag 2 identification of the color component in the data section, since the signal 1 identify common encoding/independent encoding, you can recognize that all parts of the data section in the sequence can be defined as the combined sections of the components C0, C1 and C2 (that is, each section consists of combining the information of the three color components).

The scheme of work of the second unit of the image encoding

11 illustrates the internal configuration of the second unit 7a, the image encoding. Figure 11 assumes that the input video signal 3A is entered by macroblocks, a macroblock consists of samples of the component C0 in the format shown in figure 4. The second blocks 7b, 7C, the image encoding have the same internal configuration, except for processing input video signals 3b (component C1) and 3C (component C2) instead of the input video signal 3A. Therefore, the operation of the second unit of the image encoding will be further described by means of a typical example storyblock 7a, the image encoding.

(3) the Case in which the signal 13 specifying only intracoronary indicates "not only interaktywna"

Block 115 predictions selects the reference image data from the reference image for prediction with motion compensation, stored in the memory 116, and performs the process of prediction with motion compensation for each macroblock. In the memory 116 may store multiple data elements of the reference image, each of which consists of one color component, on multiple points in time, such as the most recent points in time, or past and future points in time. Block 115 prediction performs prediction of the movement, choosing the optimal reference image for each macroblock of these reference images. It should be noted that each of the second blocks 7a-7C, the image encoding can be configured to use data from the reference image for each target color component and did not access the data of the reference image of the other color components. Therefore, the memory 116 may be configured, providing memory blocks for the three color components, respectively, and having a configuration that combines the blocks of memory in a single memory. In addition, in the sequence in which the second blocks 7a-7C encoding picture is of performing the encoding process, the first unit 5, the image encoding does not work, and therefore, the memory 116 is configured to be used in conjunction with the memory 16. There are seven types of sizes of blocks that perform prediction with motion compensation. First of all, as illustrated in figure 10(a)-10(d), choose any of the sizes 1616, 168, 816 and 88 for each macroblock. Then, when the selected size 88, as illustrated in figure 10(e)-10(h), for each block of size 88 is selected any of the sizes 88, 84, 48 and 44. In the process independent of the encoding performed by the second unit 7, the image encoding for components C0-C2 are selected and applied individual block size prediction with motion compensation.

Block 115 prediction performs the process of prediction with motion compensation on the level of macroblocks with respect to all or part of the dimensions/sizes of the sub-blocks shown in figure 10, the motion vector in a predetermined search range and one or more parts suitable for use reference images, thus producing at the output service information 117 predictions, containing the motion vector and the index of the reference picture used for prediction, and the predicted image 133. MyCitadel 118 of the predicted image 133 and the input video signal 13A is alucam for each block in the differential signal 119 predictions which serves as a single element prediction with motion compensation. The block 120 to determine the encoding mode performs a selection process based on the prediction process performed by block 115 prediction, and outputs the selected differential signal 119 predictions and type macroblock 121/type submicromolar. All the elements of the header information of the macroblock, such as the type of macroblock type submicromolar, the index of the reference image and the motion vector, defined as header information for the input video signal 3A. Then the header information of the macroblock is used for encryption and is multiplexed in the bit stream 9a. In the case of an optimality criterion of the efficiency of prediction estimated value of the prediction error with respect to the input video signal 3A serving as the target of the encoding process.

Similarly, the block 115 predictions also performs prediction with intracoronal. When the prediction with intracoronal, information about a prediction mode with intracoronal displayed on the signal 117.

Mode prediction with intracoronal in the respective color components of the input video signal 3 is selected and used individually for the respective color components. It should be noted that data from the reference image for the performance of the predictions with intracoronal involve the use of a local decoded image 123 before processing debateroom filter 22 (although it is not illustrated, the local decoded image 123 is used as data of a reference image for prediction with intracoronal, and therefore temporarily stored in block 115 predictions). As for predictions with intracoronal estimated value of the prediction error with respect to the input video signal 3A serving as the target of the encoding process. Finally, the block 120 to determine the encoding mode selects through evaluation from the point of view of efficiency predictions or coding efficiency, whether as a type of the macroblock prediction with intracoronal or prediction with internatioanal.

Block 124 conversion converts the differential signal 119 predictions and outputs the converted signal as the conversion factor for unit 125 quantization. In this case, the block size used by a single element to perform a conversion can be selected from sizes of 44 and 88. When the AVC encoding target block prediction mode prediction with intracoronal made with the possibility of adjusting the size of the transform block. When the size of the transform block is selected, the selected block size when encoding is reflected in the value of the flag 126 destination size conversion unit, and the flag 126 is multiplexed in bitove the thread 9a. Block 125 quantization produces quantization received at the input of the conversion coefficient on the basis of the parameter 127 quantization and weighting factor 12A, 12b or 12C quantization and outputs the quantized result as a quantized coefficient 128 conversion block 129 coding with variable length.

Next will be described the processing unit 125 quantization. The conversion factor is converted into a signal of a frequency domain signal from the spatial domain by block 124 conversion, is divided into the low frequency region, where the distortion can be clearly seen by the human eye, and the high-frequency region, where the distortion is difficult to detect. Thus, each frequency domain is suspended. In the low frequency range is fine quantization, while in the high frequency region is coarse quantization, whereby it may be made the quantization process, adapted to the human visual perception. Each of the weights 12A, 12b, 12C quantization is a weighting parameter that is specified in each frequency region. To convert a block of size 44 used 16 weight parameters, and to convert a block size of 88 are used 64 weight parameter. As described above, the weighting coefficients 12A, 12b, 12C, quanto the project multiplexed in a set sequence. However, in the case of the execution of the process of independent coding for the three color components can be used different weighting coefficients quantization. In line with this, all these three ratios 12A, 12b and 12C can be multiplexed, and in the case of the same values can also be multiplexed only one factor, together with information indicating that goal. Unit 125 performs quantization process weighted quantization on the transform coefficients of the three color components by using each weighting factor 12A or 12b, or 12C quantization, and thus receives the quantized coefficient 128 conversion. Quantized coefficient 128 conversion is encoded using the statistical code through unit 29 coding with variable length, using tools such as coding, Huffman and arithmetic coding.

Next, the quantized coefficient 128 conversion is restored in the local decoded differential signal 132 predictions through the block 130 inverse quantization and block 131 inverse transformation. Then the adder 134 adds the local decoded differential signal 132 with the image 133 predictions based on the selected type macroblock 121/type submicro the Loka and service information 117 forecasting, and, thus, a local decoded image 123. The local decoded image 123 after processing debateroom filter 122 is stored in memory 116 to be used for further process prediction with motion compensation. Quantized coefficient 128 conversion type macroblock 121/type submicromolar, service information 117 predictions and parameter 127 quantization, which are input to block 129 coding with variable length, are placed and executed in accordance with a predefined syntax of the encoded data of the macroblock and are packaged (AVC encoding is also referred to as packetization period of the NAL level) in the data entry section, consisting of one macroblock or merging of multiple macroblocks in the format shown in figure 4. Then the data packets are issued as a bit stream 9a.

(4) the Case in which the signal 13 specifying only intracoronary indicates only interaktywna"

Block 115 prediction performs a process of predicting only intructional described in Chapter (3). When the prediction with intracoronal, information of the prediction mode with intracoronal is issued service information 117 predictions. It should be noted that data from the reference image for prediction with INTRAC what derounian include using the local decoded image 123 before processing debateroom filter 122 (although it is not illustrated, the local decoded image 123 is used as data of a reference image for prediction with intracoronal and therefore temporarily stored in block 115 predictions). Prediction with intracoronal in the AVC encoding described above is performed with respect to an input video signal 3A. Therefore, the prediction mode with intracoronal in the respective color components of the input video signal 3 is selected and used individually for the respective color components. The block 120 to determine the encoding mode selects the mode prediction with intracoronal through evaluation from the point of view of efficiency predictions or coding efficiency.

The block 120 to determine the encoding mode outputs the selected differential signal 119 predictions on the block 124 conversion.

Block 124 conversion converts adopted a differential signal 119 predictions and outputs the converted signal as the conversion factor for unit 125 quantization. In this case, the block size used by a single element to perform a conversion can be selected from many possible dimensions, such as 44 and 88. The AVC encoding configured in such a way that the target block prediction mode prediction with intracoronal is adjusted according to the size of the block of transformed who I am. In the case where the block size conversion is done, select, selected when coding the block size is reflected in the value of the flag 126 destination size conversion unit, and the flag 126 is multiplexed in the bit stream 9a. Block 125 quantization produces quantization received at the input of the conversion coefficient on the basis of the parameter 127 quantization and weighting factor 12A quantization and outputs the quantized result as a quantized coefficient 128 conversion block 129 coding with variable length. The example process described above.

Quantized coefficient 128 conversion is encoded using the statistical code through unit 129 coding with variable length, using tools such as coding, Huffman and arithmetic coding.

Quantized coefficient 128 conversion is restored in the local decoded differential signal 132 predictions through the block 130 inverse quantization and block 131 inverse transformation. Then the adder 134 adds the signal 132 from the image 133 predictions based on the service information 117 prediction and, thus, a local decoded image 123. If information 113 specifying only intracoronary indicates only interaktywna", the prediction for what pensala movement is not performed, and, therefore, the filter 122 Troubleshooting blocking does not perform any processing or recording data as the reference image in the memory 116. With this configuration, a memory access and arithmetic operations required to process debateroom filter 22 may be reduced.

Quantized coefficient 128 conversion type macroblock 121/type submicromolar (set mode intructional), service information 117 predictions and parameter 127 quantization, which are input to block 129 coding with variable length, are placed and executed in accordance with a predefined syntax of the encoded data of the macroblock and packetizers (AVC encoding is also referred to as packetization period of the NAL level) in the data entry section, consisting of one macroblock or merging of multiple macroblocks in the format shown in figure 4. Then the data packets are issued as a bit stream 9a.

The second blocks 7a-7C, the image encoding in accordance with a first variant embodiment of the invention encode all parts of the data sections in sequence with one color component section (i.e. the section of the component C0 or section of the component C1, component or section C2) via signal 1 identify common encoding/independent encoding. Therefore, the flag is 2, the identifier is the color component multiplexed at the beginning of the data section, whereby the device for decoding can recognize which section to which data of the image corresponds to the element access. For example, the second block 7a, the image encoding sets "0" as the value of the flag 2 identification color component, the second block 7b, the image encoding sets "1" as the value of the flag 2 identification color component and the second block 7 with the image encoding sets "2" as the value of the flag 2 identification color component, and each value of the flag is attached to the beginning of the data section. Accordingly, even when the sequential multiplexing bit streams 9a-9c in the bit stream 11, the device for decoding can easily recognize which section corresponds to the encoded data components C0 or C1 or C2. In other words, the second blocks 7a-7C, the image encoding can generate the bitstream at any time, when accumulated data for one section, accumulating a corresponding output bit streams for a single image.

It should be noted that the number of macroblocks included in the section isn't limited by the value of the information 13 reference only intracoronary and information 14 about the size of the image. Information 14 about the size of the image is used as input information for the block 129 code is the simulation of variable length. Block 129 encoding with variable-length sets on the basis of information 14 about the size of the image is the upper limit of the number of macroblocks included in the section. Block 129 coding with variable length pre-counts the number of coded macroblocks, and when the number of macroblocks included in the section reaches the upper limit value, closes the data package section. Subsequent macroblocks pekititirsia how the data in the new partition. It should be noted that the information 14 about the size of the image has the same value for the components C0, C1, C2 in the case of format 4:4:4, and therefore may be enough for multiplexing only one piece of data in a set sequence.

In addition, the first unit 5, the image encoding and second blocks 7a-7C, the image encoding vary from the point of view, do they look at the header information of the macroblock as the information common to the three color components, or as the information of one color component, and the structure of the bitstream data section. In accordance with this, the block 24 conversion unit 31 inverse transform, quantization block 25, block 30 inverse quantization and deblocare filter 22 shown in Fig.9, implemented with repetition for the three color components of the arithmetic unit 124 into the project, block 131 inverse transformation unit 125 quantization unit 130 inverse quantization and deblokiruyuschee filter 122 shown figure 11. Therefore, part of the internal configuration of the first unit 5, the image encoding and second blocks 7a-7C, the image encoding can also be implemented through use of a common functional blocks. In accordance with this it is possible to implement various devices for encoding so that re-apply, for example, the same scheme several times through a suitable combination of components shown in figures 9 and 11, but not be limited to a completely independent processing block coding, as shown in Fig. In addition, as described above, if there are consecutive in the planes of the memory layout 16 in the first block 5, the image encoding, memory reference image may be used together with the first unit 5, the image encoding and second blocks 7a-7C, the image encoding.

Fig illustrates the circuit configuration of the device for decoding in accordance with the first variant embodiment. On Fig the General process of decoding is performed by the first block 302 decode the image, and the process of independent decoding is performed by block 303 to determine the color components and the second blocks 304 decoding image (provided for the three color components).

Bit stream 11 is divided into the elements of the NAL level by block 300 header analysis of high order. The header information of high order, such as a set of sequence parameters and the set of parameters of the image is decoded and stored in memory 305 header of a high order, and can be accessed the first block 302 decoding the image block 303 to determine the color components and the second blocks 304 decoding of the image. Signal 1 identify common encoding/independent encoding weights 12A-12C quantization information 13 reference only intracoronary and information 14 about the image size, which are multiplexed for each sequence, stored as part of the parameter set sequence in the memory 305 header of a high order.

Signal 1 identify common encoding/independent encoding is supplied to the switch 301. If the signal is 1 identify common encoding/independent encoding specifies the General process of coding, the switch 301 returns all elements of the NAL level sections in sequence as the bit stream 8 to the first block 302 decoding of the image. If the signal is 1 identify common encoding/independent encoding specifies the process of independent coding, the switch 301 issues useelement level NAL sections in sequence at block 303 determination of the colour components. The first and second blocks of the decoded image will be described in detail later.

Block 303 to determine the color component analyzes the value of the flag 2 identification of the color components shown in Fig.7, from the adopted level elements NAL section, then recognize which image color component in the current item match access-level elements NAL section, and distributes the elements of the NAL level of the section as bit streams 9a-9c second blocks a-s decoding of the image. As a result of this configuration of the device for decoding even when the reception of bit streams, which are inserted with alternation and coded sections for each color component in the element access, you can easily identify the section to which the color image of the component belongs, and the encoded bit stream can be properly decoded.

Scheme of work the first block decoding image

Fig shows the internal configuration of the first block 302 decoding of the image. The first block 302 decoding image takes the bit stream 11 that is output from the device for encoding, shown in Fig, for example, in the format of the bitstream 8, built up of sections, each of which consists of combining the three color components C0, C1 and C2. The first block 302 is kodirovanija image restores the output video frame by performing the decoding process for each macroblock. A macroblock consists of samples (pixels) of the three color components, shown in figure 3.

Block 310 decoding variable length takes the bitstream 8, decodes the bit stream 8 in accordance with a predetermined rule (syntax) and extracts from the bitstream 8 quantized coefficient 28 conversion for the three color components and header information of the macroblock (macroblock type 21/type submicromolar, service information 17 macroblock flag 26 destination block size of conversion and option 27 quantization), used for the three color components.

Next, the quantized coefficient 28 transformation with parameter 27 quantization is applied to the input unit 30 of the inverse quantization, which performs the same process as in the first unit 5, the image encoding and, thus, is the process of inverse quantization. In this case, the weights 12A-12C quantization used for the respective color components are applied by reference to the memory 305 header of a high order. It should be noted that if the three weighting factor 12A-12C quantization take the same value, the decoder does not have to have factors as three pieces of data, and can be shared by one data element. Next, the output and the information is sent to the input unit 31 inverse transformation, which performs the same process as in the first unit 5, the image encoding, and output information is restored in the local decoded differential signal 32 predictions (if the bit stream 8 has a flag 26 destination block size conversion, this flag 26 turn in the process of inverse quantization and the inverse process of conversion). On the other hand, block 311 predictions, referring to the service information 17 predictions in the prediction unit 15 of the first unit 5, the image encoding includes only the process of forming a predicted image 33, and the macroblock type 21/type submicromolar and proprietary information 17 predicting an input unit 311 predictions and thus the predicted image 33 for the three color components.

If the macroblock type indicates that the type of the macroblock is a prediction with intracoronal predicted image 33 for the three color components obtained from the service information 17 predictions in accordance with the information about the prediction mode with intracoronal. If the macroblock type indicates that the type of the macroblock is a prediction with internatioanal predicted image 33 for the three color components obtained from the service information 17 prediction is according to the motion vector and reference index image. The adder 34 adds the local decoded differential signal 32 predictions and the predicted image 33, thereby obtaining a temporary decoded image 323 for the three color components. Temporary decoded image 323 is used to predict the motion-compensated macroblock, and therefore, after execution of the release process in the timing of the temporary decoded image for the three color components using deblokiruyuschee filter 22, which performs the same process as in the first unit 5, the image encoding, displayed and stored as decoded image 313 in memory 312. The memory 312 stores a lot of data sets of the reference image, each of which consists of three color components for many times. Block 311 prediction generates a predicted image, selecting a reference picture indicated by the reference index images extracted from the bitstream 8 for each macroblock, from the data of the reference image. With regard to the location data of the reference image in the memory 312, these pieces of data can be stored separately or in series planes for each color component, and the samples (pixels) of the respective color components can also be stored in series points. Dec is deroanne image 313 is defined as a color video frame, containing three color component.

In addition, the first block 302 decoding of the image may be configured as follows. If information 13 reference only intracoronary stored in the memory 305 header of a high order, specifies only interaktywna", the reference image is unnecessary, since no correct process prediction with motion compensation. So goes the process performed in debateroom the filter 22, and recording a reference image in the memory 312 is not running. This configuration makes it possible to reduce memory access and arithmetic operations required to process deblokiruyuschee filtering. However, even if "only intracoronary" deblokada filter or similar filter further processing can be performed as post-processing for displaying the decoded image.

Scheme of work for the second block decoding image

Fig illustrates the internal configuration of the second unit 304 decodes the image. The second block 304 decoding image takes any of the bit streams 9a-9c, formed of the elements of the NAL level of the section of the component C0, C1 or C2, for which the allocated bit stream 11 issued from the device for encoding, shown in Fig by unit 03 determination of the colour components. The second block 304 decoding image performs the decoding process for each macroblock. A macroblock consists of samples of one color component, as shown in figure 4, thus reducing the output video frame.

Block 410 decoding variable length takes the bit stream 9, decodes the bit stream 9 in accordance with a predetermined rule (syntax) and extracts the quantized coefficient 128 conversion to a single color component and the header information of the macroblock, applied to a single color component (type 121 macroblock/type submicromolar, service information 117 macroblock flag 126 destination block size of conversion and option 127 quantization). Quantized coefficient 128 transformation with parameter 127 quantization is applied to the input unit 130 inverse quantization, which performs the same process as in the second block 7, the image encoding and, thus, is the process of inverse quantization. Since the weighting coefficient quantization used in this case, based on the flag 2 identification color component decoded by block 410 decoding variable length, weights 12A-12C quantization in memory 305 header of a high order, you select one weighting factor quantum is Oia, corresponding to the specified color component, and accessing the selected weight coefficient quantization. Next, the output unit 130 of the inverse quantization is applied to the input unit 131 inverse transform, which performs the same process as in the second block 7, the image encoding, and restored in the local decoded differential signal 132 predictions (if the bit stream 9 has a flag 126 destination block size conversion, this flag 126 treated in the process of inverse quantization and the inverse process of conversion).

On the other hand, block 411 prediction includes only the process of forming a predicted image 133, referring to the service information 117 prediction unit 115 predictions of the second unit 7, the image encoding, and takes on the type of macroblock 121/type submicromolar and service information 117 predictions, and, thus, it is predicted image 133 for one color component. If the macroblock type indicates the prediction with intracoronal predicted image 133 for one color component is obtained from the service information 117 predictions in accordance with the information about the prediction mode with intracoronal. If the macroblock type indicates that the type of the macroblock before the hat is the prediction with internatioanal, the predicted image 133 for one color component is obtained from the service information 117 predictions in accordance with the motion vector and reference index image. The adder 134 adds the local decoded differential signal 132 predictions and the predicted image 133, thereby obtaining a temporary decoded image 423 for one color component. Temporary decoded image 423 is used for subsequent prediction of the motion-compensated macroblock. Therefore, after execution of the release process in the timing of the temporary decoded image for one color component using deblokiruyuschee filter 122, which performs the same process as in the second block 7, the image encoding, the temporary decoded image 423 is output as a decoded image 413, which should be stored in the memory 412. The decoded image 413 includes samples of only one color component and, as shown in figure 5, is formed as a color video frame by connecting the corresponding output of the second blocks 304 decoding images for other color components.

In addition, the second block 304 decoding of the image may be configured as follows. If information is the situation 113 specifying only intracoronary, stored in the memory 305 header of a high order, specifies only interaktywna", the reference image is unnecessary, since no correct process prediction with motion compensation. So goes the process performed in debateroom filter 122, and recording a reference image in the memory 412 is not running. This configuration makes it possible to reduce memory access and arithmetic operations required to process deblokiruyuschee filtering. However, even if "only intracoronary" deblokada filter or similar filter further processing can be performed as post-processing for displaying the decoded image.

As can be seen from the above description, the first block 302 decoded image and the second block 304 decoding image differ from the point of view, do they look at the header information of the macroblock as the information common to the three color components, or as the information of one color component, and the structure of the bitstream data section.

Therefore, the main blocks of the decoding process, such as a block of the prediction block of the inverse transform and the block of the inverse quantization shown in Fig and 14, can be implemented by means of functional blocks common to the first bloke decoding image 302 and the second unit 304 decodes the image. In accordance with this it is possible to implement various devices for decoding by a suitable combination of the main components shown in Fig and 14, but not be limited to a completely independent unit of decoding processing, as shown in Fig. Moreover, if successive planes layout of the memory 312 in the first block 302 decoding image memory structures 312 and the memory 412 can be made common to the first block 302 decoded image and the second unit 304 decodes image 304.

The first block of the decoded image and the second block of the decoded image in the first variant of the embodiment have been described in such a way that the blocks decoding accept the bit stream 11, issued by the device for encoding. However, the bit stream applied to the input of the first block of the decoded image and the second block of the decoded image is not limited to the bit stream generated by the device for encoding, and can be supplied to the input bit streams read from the data carriers such as hard disk and digital versatile disk (DVD), and bit streams read from the server and transmitted over the network.

It should be noted that the device for encoding and decoding as described above in which a version of the embodiment of the invention in the case of the independent encoding, as shown in Fig, can achieve the same effect via the serial link of the image data for the three color components and the processing of these data elements as one set of image data. In this case, the arrangement is such that in accordance with one array of color pixels of three pieces of data are connected in a vertical direction relative to the color video signals with N pixels in the horizontal direction and V lines in the vertical direction, and for information 14 about the image size is set to N pixels in the horizontal direction and (V3) lines in the vertical direction. It should be noted that in the device for encoding and device for decoding in accordance with the first variant embodiment, the signal corresponding components could be perceived as an independent image, the correlation among the color components is eliminated in the process of encoding/decoding on the boundaries shown in bold dashed line on Fig, between the respective color components. Through the example you set the following conditions to remove the dependency.

When encoding the macroblocks located at the boundaries between the respective color components, the search for the motion vector does not use the neighboring pixels of the other C is Etowah components, and the search is conducted outside the plane to stretch the edge pixels of its color component in the same manner as in the processing of the boundary plane. When decoding the macroblocks located at the boundaries between the color components, if the motion vector deviates from the image of its color component (if you are searching outside the plane), instead of using signals of pixels of other colors predicted image is formed by stretching the edge pixels of its color component, which implied boundary plane.

Processing debateroom filter 22 is not performed between the blocks located at the boundaries between the respective color components.

- In the case of a student arithmetic coding process coding with variable length/decoding with variable length in the processes of encoding/decoding macroblocks of the color components, individually for each color component is provided by the probabilistic model, and the learning process is performed independently for each color component.

The independence of the signal of each color component is obtained by applying these conditions, and the first, second and third blocks of coding/decoding image can issue LNAT processing independently.

In addition, introduced such restrictions, which prohibited the definition of the partition through the boundaries of the respective color components is not allowed in one section of the encoded data of several different color components and the initial data of the macroblock of each color component invariably become the initial macroblock data section.

Configuration to determine what data section what color components belong, may contain an explicit definition of the color component to which the section belongs, by definition flag color_channel_idc" identification color component and attaching the flag to the initial data field section. Another configuration may contain the use of non-flag color_channel_idc", a address of the starting macroblock of each partition data and information 14 about the size of the image, and recognizing what data section which color component prinadlezhashtimi, when the number of horizontal pixels is W=1920 and the number of vertical pixels is V=1080, "0", "8160" and "16320" is set as the address of the starting macroblock of the components C0, C1 and C2 nodes, then the macroblock with address "0-8159" selected for the component C0, the macroblock with address "8160-16319" dedicated to component C1 and the macroblock with address "16320-24479" selected for comp is the component C2.

With this configuration, the structure of the image/element access for a process common encoding/independent encoding can be made common, thereby increasing the efficiency of operations for random access and editing.

The method of decoding image decoding a color image based on the input bit streams generated by running over a colored image in the format of 4:4:4 encoding with compression, encoding, compression is performed by transforming and quantizing the color image at the block level, the method of decoding image contains:
stage decoding for decoding the first identification information indicates that the coded whether the signals of respective color components independently, and to decode the second identification information when the first identification information indicates that the signals of respective color components are encoded independently, the second identification information indicates whether all color components of the intra-coded, and
in the case where the first identification information indicates that the signals of respective color components are encoded independently, and the second identification information indicates that all from the expressions of the respective color components are intra-coded, stage decoding generates a decoded image, in which we disable execution deblokiruyuschee filtering at each boundary between blocks that serve as elements of the transform and quantization.



 

Same patents:

FIELD: information technology.

SUBSTANCE: image processing circuit corrects colour in a predetermined colour range so that it is corrected such that it partially includes the reference range of red hue, lying in the centre of the colour range of the constant hue, lying from the achromatic colour which is the colour of the lowest saturation in the extended colour reproduction region, to the red colour having the highest colour saturation in the extended colour reproduction range, but does not include red colour having the highest colour saturation in the extended colour reproduction range. At that moment, the image processing circuit corrects said colour such that its hue is replaced with hue lying close to the yellow colour in the extended colour reproduction range, and hue from the reference range of red hue is replaced with red colour hue in the sRGB standard colour reproduction range, and further, the value by which the hue changes to hue lying close to the yellow colour is less than the colour lying further from the reference range of red hue in the extended colour reproduction range than for the colour lying close to the reference range of red hue in the extended colour reproduction range.

EFFECT: providing a video display device capable of displaying video using a display with a wide colour gamma based on a video signal corresponding to a standard having a narrower colour reproduction range than the range of the display with a wide colour gamma, while solving the problem of hue shift when displaying red colour with average saturation.

3 cl, 9 dwg

FIELD: information technology.

SUBSTANCE: in an image encoding system compression processing is applied to an input image signal, comprising multiple colour components, encoded data obtained after independent encoding processing of the input image signal for each of the colour components, and the parameter that indicates which colour component corresponds to encoded data is multiplexed with the bit stream.

EFFECT: higher encoding efficiency and providing possibility to include data for one image in one access unit, and establish identical time information and single encoding mode for corresponding colour components.

6 cl, 25 dwg

FIELD: information technology.

SUBSTANCE: when controlling a diffused illumination element, the category of data displayed by the unit is identified. Diffused illumination data associated with the identified category are extracted and the extracted diffused illumination data are displayed according to the displayed data. The extracted diffused illumination data can be a diffused illumination script which can determine temporary parts of the diffused illumination data. Diffused illumination data can be associated with a category based on user input. A data subcategory can be identified and diffused illumination data can be modified with additional diffused illumination data associated with the subcategory. Association of the category with diffused illumination data can be edited by the user. Default association of the category with diffused illumination data can be provided.

EFFECT: eliminating the direct link between the diffused illumination effect and context of the displayed video.

18 cl, 3 dwg

FIELD: information technology.

SUBSTANCE: secondary video signal is generated, said signal being composed of signals having values derived via conversion of intermediate values into values lying inside an output range according to a predetermined conversion rule when intermediate brightness values (determined by formulas where Smin is the output value of the lower limit, Xr to Xb are values of RGB signals of the main video signal, k is a constant, and Lr to Lb are intermediate values of RGB brightness), include a value greater than the value of the upper output limit, otherwise a secondary video signal consisting of a signal having an intermediate brightness value is generated.

EFFECT: preventing gradation error when a given video signal shows colour in a region outside the colour range of the video display element, performing signal conversion processing with low arithmetic load.

16 cl, 7 dwg

FIELD: information technology.

SUBSTANCE: scalable video codec converts lower bit depth video data to higher bit depth video data using decoded lower bit depth video data for tone mapping and tone mapping derivation. The conversion can also be used for filtered lower bit depth video data for tone mapping and tone mapping derivation.

EFFECT: high encoding efficiency.

7 cl, 3 dwg

FIELD: physics.

SUBSTANCE: when controlling an ambient illumination element, a host event is detected, a light script associated with the detected event is retrieved and the retrieved light script is rendered in accordance with the detected event. A user may associate the light script with the event and/or an event type which corresponds to the event. A default association of events and/or event types may be provided, although these default associations can be modified by the user. An event type which corresponds to the event can be identified and a light script associated with the identified event type can be rendered in response to the detected event.

EFFECT: reduced viewer fatigue and improved realism and depth of experience.

20 cl, 3 dwg

FIELD: information technology.

SUBSTANCE: image encoder includes the following: a predicted-image generating unit that generates a predicted image in accordance with a plurality of prediction modes indicating the predicted-image generating method; a prediction-mode judging unit that evaluates prediction efficiency of a predicted image output from the predicted-image generating unit to judge a predetermined prediction mode; and an encoding unit that subjects the output signal of the prediction-mode judging unit to variable-length encoding. The prediction-mode judging unit judges, on the basis of a predetermined control signal, which one of a common prediction mode and a separate prediction mode is used for respective colour components forming the input image signal, and multiplexes information on the control signal on a bit stream.

EFFECT: high optimality of encoding the signal of a moving image.

4 cl, 86 dwg

FIELD: information technology.

SUBSTANCE: image includes, when applying encoding processing to three colour components using the 4:0:0 format, data for one image into one access module, which enables to establish the same time information or identically established encoding modes for corresponding colour components. In an image encoding system for applying compression processing to an input image signal, comprising multiple colour components, encoded data obtained after independent encoding processing of the input image signal for each of the colour components, and the parameter that indicates which colour component corresponds to encoded data is multiplexed with the bit stream.

EFFECT: high encoding efficiency owing to use of a single encoding mode for corresponding colour components.

2 cl, 25 dwg

FIELD: information technology.

SUBSTANCE: invention relates to an image signal processing device, which enables to reproduce the appearance of an image on a plasma display panel (PDP), using other display devices such as a cathode-ray tube or liquid-crystal display (LCD), while processing signals. In an image processing module, such processing for an image signal for which an image received when the image signal is displayed in a display device of another type besides a PDP, may seem like an image displayed on a PDP. At least one reproduction colour shift is performed, which is associated with a moving image which forms as a result of that, RGB glow is included in the said order of reproduction, smoothing structure used in the direction of space, reproduction of the smoothing structure used in the direction of reproduction time, the interval between pixels, and reproduction of an array of strips. The invention can be used when, for example, an image which must look like an image displayed on a PDP, is displayed on an LCD.

EFFECT: possibility of obtaining a type of image on a plasma panel display reproduced on another display different from the plasma panel display such as a liquid-crystal display, while processing signals.

6 cl, 20 dwg

FIELD: information technologies.

SUBSTANCE: it is suggested to do coding and decoding uniformly for multiple colouration formats. Based on control signal providing for type of colouration format of inlet signal from dynamic image, if colouration format is 4:2:0 or 4:2:2, the first unit of prediction mode detection with intra-coding and the first unit of predication image coding with intra-coding are applied to component of dynamic image inlet signal colouration component, and the second unit of prediction mode detection with intra-coding and the second unit of prediction image formation with intra-coding are applied to colouration component. If colouration format is 4:4:4, the first unit of prediction mode detection with intra-coding and the first unit of prediction image formation with intra-coding are applied to all colour components to do coding, and unit of coding with alternating length multiplexes control signal as data of coding, which should be applied to element of dynamic image sequence in bitstream.

EFFECT: improved mutual compatibility between coded video data of various colouration formats.

12 cl, 24 dwg

FIELD: information technologies.

SUBSTANCE: video recording device comprises a compression facility, which compresses video data in process of speed control implementation, including variation of a quantisation step so that the bit transfer speed is reduced within the previously determined period to the target speed of bit transfer, a recording facility, which records compressed data onto a recording medium, and a facility for calculation of full speed, which calculates full speed of transferring bits of the result of video data compression with the help of the compression facility from the start of compression to the current moment. The compression facility comprises a facility for limitation of a quantisation step, which limits the varied quantisation step previously for a certain upper limit, which is lower than the upper limit varied in the compression facility, and a facility for fixation of the quantisation step, which fixes the quantisation step to the previously determined upper limit, when the calculated full speed of bit transfer exceeds the target speed of bit transfer.

EFFECT: compression of video data so that bit transfer speed is reduced to target bit transfer speed together with saving minimum quality of an image, and recording of compressed video data.

6 cl, 3 dwg

FIELD: radio engineering, communication.

SUBSTANCE: digital multimedia coder/decoder is proposed, which uses a method of flexible quantisation, making it possible to vary quantisation by different measurements of coded digital multimedia data, including spatial channels, channels of frequency subranges and colour channels. A codec efficiently applies an alarm circuit to alarm various shifts of flexible quantisation combinations for initial usage scenarios. When selection of a quantiser is available, the codec efficiently codes the current quantiser, determining a subset of quantisers, and indexes the current quantizer from the set.

EFFECT: provision of flexible quantisation by different measurements of coded digital multimedia data.

18 cl, 15 dwg

FIELD: physics.

SUBSTANCE: invention relates to digital television and particularly to compression of a digital video stream in a television communication channel. A field sequence is divided into groups of three types: O-fields, which act as the reference; N-even fields, which are encoded via prediction based on the previous field within one frame; M-odd fields, which are coded with prediction based on the previous odd field from another frame. To improve visual quality of the image according to the disclosed method, odd and even fields change places in the entire video sequence or group of frames. Frames with a higher image definition are formed as a result. To increase efficiency of compressing a digital stream, the disclosed method employs cutting of readings in N-fields horizontally and vertically with subsequent restoration thereof at the receiving side, wherein two neighbouring N-fields form a full half-frame (field).

EFFECT: communication channel bandwidth reduction, as well as high efficiency of digital conversion of a video signal, lying in additional cutting of the volume of digital information and increasing visual quality of the image, with sufficiently easy implementation.

13 dwg, 1 tbl

FIELD: information technology.

SUBSTANCE: disclosed is an apparatus for encoding video and a corresponding method of applying orthogonal transformation to a prediction error signal between a video signal for a target encoding region and a predicted signal for the video signal, and sampling the obtained orthogonal transformation coefficient using a given sampling interval so as encode the coefficient. Prediction error power is calculated, which is the prediction error signal power. For input information such as the calculated prediction error power, the given sampling interval and the upper limit for the volume of the code generated for the target encoding region, it is determined whether the upper limit of the volume of the code, generated when performing sampling using a predetermined sampling interval, is exceeded or not, and the encoding process is changed based on the determination result.

EFFECT: providing a video encoding device which does not require repeated encoding or encoding which corresponds to two or more encoding modes, and performs encoding, the volume of the generated code of which does not exceed the upper limit, without waiting for the result of measuring the volume of the generated code.

5 cl, 20 dwg

FIELD: information technology.

SUBSTANCE: prediction error signals is calculated by determining values of the difference between sample values of a predicted data block and sample values of the initial input block. A modified error prediction signal is generated by replacing difference values which are overshoots with difference values which are not overshoots, wherein the difference value which is an overshoot depends on the expected value of the amplitude of the difference value. Further, transform coding is performed over the prediction error signal in order to create a first presentation of the first error signal component and spatial coding is performed over the prediction error signal in order to create a second presentation of the first prediction error signal component. The first and second presentations are merged and coded.

EFFECT: efficient spatial presentation of those prediction error signals of the same image block which are not well correlated with basic functions of the applied transformation.

33 cl, 11 dwg

FIELD: information technology.

SUBSTANCE: disclosed is an improved system and a method for providing improved inter-layer prediction for extended spatial scalability in video coding, as well as improved inter-layer prediction for motion vectors in the case of extended spatial scalability. In various versions, for the prediction of macroblock mode, the actual reference frame index and motion vectors from the base layer are used in determining the need to merge two blocks. Additionally, multiple representative pixels in a 4x4 block can be used to represent each 4x4 block in a virtual base layer macroblock. The partition and motion vector information for the relevant block in the virtual base layer macroblock can be derived from all of the partition information and motion vectors of those 4x4 blocks.

EFFECT: less complex calculations and high efficiency of coding and decoding during scalable video coding.

24 cl, 12 dwg

FIELD: information technology.

SUBSTANCE: method involves evaluating a history of transform coefficient values associated with one or more previous layers of the SVC scheme, and estimating one or more refinement coefficient values associated with a current layer of the SVC scheme based on the history. On the encoding side, the coding process may include excluding information for one or more refinement coefficient values from the bitstream and signaling to the decoder that such information is excluded from the bitstream. On the decoding side, the coding process includes parsing the bitstream to identify information which signals to the decoder that information is excluded from the bitstream, and generating such information based on the history associated with one or more previous layers of the SVC scheme.

EFFECT: high efficiency of scalable video coding based on history of history of transform coefficient values.

25 cl, 11 dwg

FIELD: information technologies.

SUBSTANCE: method is proposed for coding of clarification coefficients in layer of scalable increase of signal-noise ratio (SNR) quality in compressed sequence of video frames. Sequence of video frames is received. Forecasting of initial video signal is composed of video sequence in current frame. Differential signal is generated by subtraction of initial video signal forecast from initial video signal in current frame. Conversion is applied to differential signal. Multiple conversion ratios are quantised. Clarification coefficient is displayed into ternary clarification symbol. Clarification symbols are grouped in a certain sequence of coding. Groups of clarification symbols are coded using codes of alternate length.

EFFECT: improved efficiency of video data coding.

30 cl, 15 dwg

FIELD: information technologies.

SUBSTANCE: video sequence is processed into combination of frames, besides each frame of multiple frames is processed into multiple macroblocks. Forecast of initial video signal, which is part of macroblock in current frame, is generated from video sequence. Differential signal is generated by subtraction of initial video signal forecast from original video signal in current frame. Conversion of initial signal is applied, besides multiple conversion ratios are quantised, symbol of at least one syntactic element, which determines characteristic of differential signal, is established. Symbols of at least one syntactic element of one and the same category are coded jointly.

EFFECT: increased efficiency of information coding and decoding.

21 cl, 11 dwg

FIELD: information technologies.

SUBSTANCE: device is proposed for coding of dynamic image to form bitstream by means of arrangement of dynamic coding image with compression over digital signal, and compression is realised by means of conversion and quantising of digital signal of dynamic image on the basis of blocks, besides device comprises the following components: coding unit to multiplex identification information in bitstream, which indicates whether it is required to intra-code all images that comply with the digital signal of dynamic image, and indicates availability or absence of blockiness elimination filter at each border between the blocks serving as single elements of conversion and quantising.

EFFECT: development of method for formation of bitstream to provide for compatibility between bitstreams.

4 cl, 23 dwg

FIELD: physics, communication.

SUBSTANCE: invention relates to picture digital signal encoder/decoder in appropriate chroma signal format. Encoder comprises module to divide input bit flow into appropriate colour components, module to divide colour component input signal into blocks so that coding unit area signal can be generated, module to generate P-frame for said signal, module to determine prediction mode used for encoding by efficiency P-frame prediction, module to encode prediction error for encoding difference between predicted frame corresponding to prediction mode as defined by appropriate module and colour component input signal, and module for encoding with variable length of prediction mode code, output signal of prediction error encoding module and colour component identification flag indicating colour component wherein belongs said input bit flow resulted from division of colour components.

EFFECT: higher efficiency.

2 cl, 98 dwg

Up!