Extrapolation of movement vector for video sequence code conversion

FIELD: converting code of received video sequence using extrapolated movement data received from video sequence.

SUBSTANCE: proposed method for code conversion involves reception of first bit stream of compressed picture data having some coding parameters. These parameters may relate to GF-structure of picture frames, picture frame size, to parameter showing if frames presented in input bit stream are picture fields or frames, and/or if they form picture frames presented in bit stream, direct or interlaced sequence. First and second movement vectors are obtained from input bit stream and used together with weighting coefficients to extrapolate third movement vector for output bit stream of compressed picture data. Output bit stream that differs from input one by one or more parameters is outputted as converted-code output signal.

EFFECT: provision for minimizing of or dispensing with movement estimation in code conversion process.

22 cl, 4 dwg, 1 tbl

 

The United States government has a paid-up license in this invention and the right in certain circumstances to require the patent owner to license others within a reasonable period stipulated by the terms of contract number 70NANB5H1171 issued by the National Institute of standards and technology.

The technical field to which the invention relates.

The present invention relates to motion estimation during encoding conversion code sequence and, in particular, to a method for transcoding of the received sequence using the extrapolated information movement from the received sequence, which allows to minimize or eliminate the motion estimation process of the conversion.

The level of technology

Methods of data compression allow you to transfer large amounts of data in relatively narrow frequency bands. The algorithms used in the compression system, depend on the available bandwidth or memory capacity, features arising from the application and capabilities of the hardware required for the implementation of the compression algorithm (encoder and decoder). A standard for compressing moving picture experts group moving pictures of the type MPEG 2 (MPEG2), given here as a reference, is a well-known ways is Ohm video compression. Motion estimation, which is the process that is used in the encoders to calculate motion vectors, usually considered the most expensive stage of the encoding process. Similarly, the motion estimation is likely to be the most expensive step in the process of conversion, in which the sequence decode and then encode again with the new settings. With the advent of high-definition television (HDTV) this point is of particular importance, because television studios have to recode the bit streams of MPEG-2 human data from one format to another. For example, television Studio, working in a standard HDTV, you should be able to encode the bit stream of MPEG-2 from the same group structure image frames (GK-structure) to another, from one frame size to another and/or from one bit rate to another. There are cases when these studios have to transcode the footage to MPEG-2 from the field images in full frames of images or frames of images in the fields of images. There are cases when the studios need to transcode the footage to MPEG-2 alternating sequence in a direct sequence or from a direct sequence in alternating sequence.

Transcoding can also be used for the sushestvennee function “trim”, the input signal which represents the main image, and the encoded output signal represents the portion of the image in the main image.

The invention

Thus, the purpose of the present invention is to develop a method of transcoding a video sequence, which allows to minimize or eliminate motion estimation.

Another goal is to develop a method of transcoding a video sequence, according to which the information of the movement of the input bit stream is used to generate the motion vector for the encoded output bit stream.

Another goal is to develop a method of motion estimation for the process of conversion, which supports all the prediction modes MPEG-2.

The principles of the present invention allow to achieve these and other objectives through the development of a method of transcoding according to which receive the first bit stream of compressed image data having the same or different encoding options. These parameters can relate to a GK-structure of frames of images represented by the first bit stream, the size of frames of images represented by the first bit stream, whether the frames are presented in the first bit stream, the fields on images or full frames of images, and/or the will is formed whether frames of images, presented in the first bit stream, a direct or alternating sequence. From the first bitstream receive the first information movement and use it to extrapolate the second movement of the second bitstream of compressed image data. As a re-encoded output signal to output a second bit stream, different from the first bit stream according to one or more parameters.

Brief description of drawings

To fully appreciate the invention and numerous additional advantages are to be treated to the following detailed description, given with reference to the accompanying drawings, on which:

figure 1 illustrates the basic principles of extrapolation of the motion vectors

figure 2 illustrates the system of recoding, constructed in accordance with the principles of the present invention,

figure 3 illustrates the process of encoding image frames from one Ledger structure to another in accordance with the principles of the present invention, and

figure 4 illustrates the General case of extrapolation of the motion vector and can be used to better understand the extrapolation of the motion vector applied to the conversion.

Detailed description of the invention

Figure 1 illustrates the basic principles of extrapolation of the motion vector. In General SL the tea extrapolation of the motion vector is a method of motion estimation for image sequence, when somewhere in the sequence of movement is known. In figure 1, the square on the image, selected a thicker line represents a single block of pixels (according to the terminology of the American Association communication (AAC), absorbing) in the video frame images. It is assumed that each block in the frame at a fixed position. According to MPEG-2, these blocks are referred to as “macroblocks”, and each of them consists of a block of 16x16 pixels. A solid arrow in figure 1 shows the known movement of the block allocated thickened line, as the block moves from frame images in the frame image. In fact, move the image in the block, not the block, occupying a fixed position. This is a known movement specified by the motion vector, is used to extrapolate the motion of the block when it moves from frame To frame S. the Dotted arrow in figure 1 indicated the extrapolated motion vector. The arrows in figure 1, in fact, point in the opposite direction to the movement, as figure 1 (as well as other accompanying drawings) use more General kopiowanie agreement, according to which the vector associated with a block indicates the “source” of this block in the reference frame image. In practice this motion vectors present in the form of a set of coordinates x and y, which, with therefore, its, indicate the horizontal and vertical movement of the block between two frames of images. When extrapolation is usually accepted that the movement is homogeneous in time and smooth in space. Accordingly, it is expected that the block in the frame will move to the position shown in frame With, and it is expected that all overlapped the blocks have approximately the same motion.

Figure 2 shows the system of recoding constructed according to the principles of the present invention. According to figure 2, the system 200 transcoding takes the input bitstream 201 of the image data compressed according to the standard MPEG-2. This input bitstream 201, which is characterized by a certain bit rate, represents the frames of this size and General Ledger structure. Personnel data input bit stream 201 can consist of fields of the image or full image frames and may be direct or alternating sequence. MPEG-2 decoder 2, known from the prior art, receives the input bit stream 201 and decodes it to generate the output signal in the form of the decompressed digital image data 203. The decoder 202 also outputs information of the motion (i.e. motion vectors) from the input bit stream 201 as a separate output signal 204. Digital image data 203 is coming to the module ITE processing, who works at the pixel domain and is capable of well-known way to change the size of frames of images represented digital image data 203. In addition to changing the frame size, the intermediate processing module 205 may perform known cutting function, isolating the portion of the image in the main image being rendered digital image data 203, and generates an output signal in the form of data representing a portion of an image. The intermediate processing module 205 is also capable of well-known operations of rotation and removal of alternation, which is required when the direct conversion of a sequence of frames in alternating sequence of frames when encoding interlaced sequence of frames in a direct sequence of frames. Module 205 intermediate processing operates on the external input signal 206, which indicates which (if any) the processing operation to be accomplished and also specifies the parameters that shall be used in such processing. Considering the fact that the operations module 205 intermediate processing are optional, if the image resizing, cropping, rotation and removal of the rotation of the frame in this embodiment, application of the conversion is not carried out, the module 205 intermediate processing works is just like a relay module and generates an output signal in the form of digital image data 207 without any processing of the pixel domain. In this case, the digital image data 207 and the digital image data 203, obviously, are the same. Alternatively, if this option is application of conversion provides for the implementation of one or more of the above-mentioned processing operations, the module 205 intermediate processing performs the operation of processing the pixel domain and generates an output signal in the form of processed digital image data 207. Digital image data 207, processed or not processed on the module 205 intermediate processing, proceed to MPEG-2-encoder 208, which re-encodes the received image data 207 with the new settings specified by an external input signal 209. For example, the encoder 208 may encode the digital image data 207 to represent frames with a GK-structure of images, other than GK-structure image frames that represent the input bit stream 201, and/or may adjust the bit rate. The encoder 208 may also be encoded digital image data 207 in the form of fields or frames. For the implementation of the encoding process, the encoder 208 uses the information of the motion contained in the output signal 204 of the decoder 202, with the aim of extrapolating motion vectors for the re-encoded frames. After that, re-encoded frames are removed from the system the volumes of transcoding 200 as the output bit stream 210, consisting of digital image data compressed according to MPEG-2.

According to the above, this output bit stream 210 can represent the frame size and GK-structure image frames other than size and GK-structure of the input bit stream 201, and the bit rate of the output bit stream 210 can be controlled so that it differed from the bit rate of the input bitstream 201. In addition, the output bit stream 210 may represent fields, whereas the input bit stream 201 is full frames, or the output bit stream 210 can represent the full frames, while the input bitstream 201 represents the fields. Similarly, the output bit stream 210 may represent alternating sequence, while the input bit stream 201 is a straight sequence or the output bit stream 210 may represent a direct sequence, while the input bit stream 201 is alternating sequence. In addition, the output bit stream 210 can represent a “cropped” version of the input bit stream 201. In essence, the system 200 transcoding takes the input bit stream 201, with certain parameters, and re-encodes the input bit stream 201 to generate the output bit stream 210 that is different from the one bit stream 201 one or more parameters. This system 200 transcoding is unique in that it uses the information of the movement of the input bit stream 201 to extrapolate information of motion for the output bit stream 210.

Figure 3 illustrates the process of conversion of images from one GK-structure image frames to another, according to the principles of the present invention. According to figure 3, the input bit stream is frames with GK-structure consisting of an independently coded frame (I-frame), frame encoded with bi-directional prediction (B-frames”), In - frame and frame encoded with prediction (“P-frame”).

Thus, the input bit stream is GK-structure “IBBP”. According to figure 3, two consecutive In-frame input bit stream is indicated for clarity as B1 and B2. This input bit stream similar to the input bit stream 201 in figure 2. During conversion, the input bit stream having a GK-structure “IBBP”, is converted into an output bit stream having a GK-structure “IPPP”. According to figure 3, three successive P-frame of the output bit stream are indicated for clarity P1, P2 and P3. This output bit stream similar to the output bit-stream 210 of figure 2. In order not to clutter the drawing, figure 3 does not depict the macroblocks that make up the individual frames of images. For about the westline process of recoding, the motion vectors of the I-P and P-B2 from the input bit stream is used to extrapolate the motion vectors for the output bitstream. For example, according to figure 3, the motion vector of the I-P from the input bit stream can be used to estimate the motion vector P1-P2 for the output bitstream. Similarly, the motion vector of the P-B2 from the input bit stream can be used to estimate the motion vector P2-P3 for the output bit stream.

Now consider the process of extrapolating motion vectors in more detail with reference to figure 4, which shows four frames of the video sequence. In principle, these four frame will be present in the output bit of the syrup, as in the input bit stream, and as usually happens in practice.

In any case, in figure 4 the frame T is “end” frame whose motion relative to the frame R to be determined. Both of these frames will be present in the output bit stream. Frames KR and KT are frames relative movement which is known from the input bit stream. According to the principles of the present invention, a known motion between the frames KR and CT can be used to extrapolate the motion vectors for the output bitstream. Frames KR and KT are referred to as “base pair”, and the frames R and T are referred to as “current pair. For motion estimation of the current pair, you can use the number of base pairs. In General, for this purpose, any potentially suitable base pair that is closest in time to the current pair.

During the conversion, it so happens that some of the blocks in frame T do not accept information extrapolation of motion. This means that it may happen that some of the blocks in frame T do not overlap movement no blocks in the surrounding base pairs. Such blocks in frame T can simply be independently encode or to calculate their movement accepted way. Alternatively, the movement of such blocks can be displayed or interpolate from nearby blocks. However, in most cases, several blocks in the base pairs will overlap the majority of blocks in frame T, and then must choose the best motion vector among several suitable motion vectors. Each overlapping block of the base pair provides one suitable motion vector. When choosing from a variety of suitable motion vectors useful to compare each motion vector weight coefficient (weight) and select the motion vector with the highest weight. Using as example the frames presented on figure 4, the weighting factor for each suitable motion vector (MV) can be calculated as follows:

Total weight (MV)=(the weight of the slab)·(Weight ratio)of·(Weight rounding), (1)

where is the Weight of a feather is rite=(the number of pixels in the block frame CT, which overlap the pixels in the corresponding block of the frame T) (2)

Weight frame=I/[1+abs (Temporary weight)], (3)

where Temporary weight=(tKT-tKR)·[(tKR+tKT)-(tT+tR)] (4)

and

Weight rounding=[1-(rounding error horizontally)]·[1 - (rounding error vertical)] (5)

In the above equations (1) - (5) tKR, tKT, tR and tT denote the points in time of the display frame, respectively, KR, KT, R, and So the Notation · and abs are, respectively, the multiplication operator and the operator of taking the absolute value. Rounding errors vertically and horizontally occur when extrapolated vector rounded up to the nearest polwechsel, and, accordingly, each of which takes values from zero (0) to one half (1/2). Although equations (1)-(5) are applied to the frames presented on figure 4, it is clear that these General equations can be applied to other human resource configurations.

In addition to simple way of selecting suitable motion vector having the highest weighting factor, the present invention provides for other ways to get the best(their) vector(s) of motion for this unit. According to alternative implementations, one can calculate the weighted average, where weights are calculated using the above equations (1)-(5). Using these weights can calculate the ü the best motion vector component by multiplying the weighting coefficients on the x - and y-components of the respective motion vectors, with the purpose of generating weighted components, the summation of the weighted components, and dividing the sum of the weighted components of the sum of the weights.

The above-described method, the weighted average can also be accomplished by “clusters”. This means that if a vector diagram of suitable motion vectors formed more than one cluster (i.e. a closed group) motion vectors, it is possible to calculate the best motion vector for each cluster. This will be the centroid or center of mass of the cluster. You can then make the final selection of the best motion vector of the best vectors of individual clusters.

It should be noted that the best motion vector can also choose among the set of available motion vectors, not using weights. For example, to determine the best motion vector is possible to calculate the root mean square error (RMSE) or mean absolute deviation (CAO), representing pixel differences between blocks. Use of estimates RMSE and CAO, of course, well known to specialists in this field of technology.

The following algorithm is presented in table 1, comprises the steps of finding the best motion vector for each possible mode predictions of the end frame (e.g. frame T pofig). These stages are represented by the pseudo-code can be programmed in any programming language, known to specialists in this field.

Table 1 pseudo-Code for finding the best motion vector.

for each prediction mode, subject to review for the final frame (the current pair),

- for each field or frame of the applicable mode (the top field, bottom field, frame)

to initialize the table, indexed by the blocks in the target frame,

for each base pair, a recognized suitable for the current pair (e.g., base pair is usually considered “qualified”if it is near time within a predetermined range),

for each motion vector of each of the independently coded block in the target frame base pairs

- to determine where the vector moves the block in the target frame,

to calculate the weighting factor vector or to evaluate the degree of coincidence (for example, RMS, CAO),

- to save the information in the table cells corresponding blocks of the final frame,

for each block in the target frame,

- to determine the best motion vector for a field or frame,

for each block in the target frame,

- to determine the best prediction mode and the corresponding(e) vector(s) of motion.

Note that some modes preds the marks, for example, predictions of field for full-frame and two strokes for P-frames, with the data block associated multiple motion vectors. In addition, for b-frames, the encoder must decide for each block, whether to use direct prediction, backward prediction, or both predictions. In some cases acceptable results do not give any of the prediction modes. This occurs when there is no suitable motion vector or when the best motion vector determined according to one of the above ways, not suitable for this application. In addition, there are cases where the encoder can simply use the motion vectors (without changing them) from the input sequence to the output sequence.

Figure 4 shows the situation when for full-frame use the prediction of the full frame, without changing the size. Even with the block in the frame of the CT associated vector vK, then the vector v associated with the one or more blocks in frame T, will have the following form:

v=Tv·vK (6)

where Tv=(tT-tR)/(tKT-tKR) (7)

Here, Tv is a time scale factor for vectors, a tKR, tKT, tR and tT denote the points in time of the display, respectively, personnel KR, KT, R, and So In the case of full frames, fields which display in different moments of time, there is neop udalennosti in the time display. In such cases, you need to take the average time on the relevant fields. To track the movement of a block from the frame CT in frame T, consider a point in the upper left corner of the individual unit. With reference to figure 4, this is the point qK in the frame CT and the point q in the frame So the movement of the point q is defined as follows:

g=qK-(Tb·vK), (8)

where Tb=(tT-tKT)/(tKT-tKR) (9)

In the process of recoding, modifying the frame size, frames KR and KT will not have the same size as the frames R and So the Change of size is calculated as follows. Let Rx and Ry represent the coefficients increase frames, respectively, vertically and horizontally.

This means:

Rx=(width of frame T)/(width of frame CT) (10)

and

Ry=(frame height T)/(frame height CT) (11)

With regard to resize frames equations (6) and (8) take the form:

v=R(Tv·vK) (12)

and

q=R(qK-(Tb·vK)) (13)

In the above equations used the notation:

R(V)=(Vx·Rx, Vy·Ry), (14)

where V is a point (pixel) or a vector with two components. Note that R refers to the different frame sizes, but not to different dimensions of the block.

Resizing is used not only to frames resized, but also to the relevant fields and complete the nomination and to the corresponding predictions of the half-frame and full frame. To take account of forecast what their fields (in relation to the half-frame and full frame),note, that the lower fields at 0.5 pixel lower than indicate their coordinates. The displacement vector, dX, half frame or full frame set as follows:

dX=(0,0), if X is the upper half frame or full frame (15)

and

dX=(0,05), if X is the bottom field. (16)

Applying this to equations (12) and (13), we obtain the new equation for v and q:

v=RTv[vK+dKR-dKT)+dT-dR (17)

and

q=R(qK+dKT-[Tb(vK+dKR-dKT)])-dT (18)

Finally, let the module 205 intermediate processing shown in figure 2, cuts the image. The above equations can be applied in the following way. Put that frames R and T have the same size as the source, while the output images embedded in them in the proper place. This place sets the region of interest to us. For extrapolation used only those blocks that overlap the region of interest to us. Then 1-4 derived from these equations apply to this case.

The above statement could give the impression of homogeneity: if frame T is full frame, using the prediction based on the full frame, the frame CT is also full frame, using the prediction-based fields. However, the above formulas are more General, as evidenced by the examples below.

Let the frame T uses prediction based on the full-frame, and the frame CT uses the etc is skazanie-based fields. According to the present invention, whether the frame CT-frame or full frame. In any case, the motion vectors of the frame of the CT-based fields, so that Ry=2 and Rx=1. If the frame CT is full frame, the blocks are 16×8; however, it is clear from the above mathematical formulas. If the unit uses the prediction 16×8, the two halves should be treated as two separate blocks.

Suppose now that the frame T and frame CT uses a prediction-based fields. In this case, since the vectors in both frames are images of the same size, resizing is not required, and Rx=Ry=l. Thus, when implementing the present invention for the equations does not matter whether the combination of the half-frames and full frames or predictions of the half-frame and full frame.

As is clear from the above explanation, if the block in the frame CT uses two stroke, two vectors are applied separately, possibly with different weights. In addition, estimation of the frame T for the two strokes is very reminiscent of his assessment for bidirectional interpolation. That is, each vector is evaluated separately, and then estimate their combination. Accordingly, the present invention supports all of the prediction modes MPEG-2: forecast of what their full frame, fields and two stroke for full-frame; and a prediction frame, two stroke and 16×8 for the fields.

Note that these equations are, in General, applicable to uniform rectilinear motion, and not to the uneven movement. Erratic movement may be caused, for example, by camera shake or constant acceleration of objects in frames. In such cases, non-uniform motion you out as follows. As for the non-uniform motion associated with camera shake, it is useful first to make brackets for General movement, and then compare the movement of blocks. With this purpose you can use the well-known calculations of RMS and CAO. As for the movements associated with constant acceleration of objects, then this movement can be extrapolated using not two, but three frames.

Although we have illustrated and described the embodiments of the present invention, considered as predominant, specialists in this field it is clear that it contemplates various changes and modifications, and replacement of its elements equivalent elements, not beyond the true scope of the present invention. In addition, it allows numerous modifications are not contrary to his nature. Therefore, it is envisaged that the infusion is her invention is not limited to the individual variants of its implementation, presented here as the best mode for carrying out the invention and that the present invention includes all the embodiments covered by the scope of patent protection of the attached claims.

1. The method of encoding image data, according to which receive the input bitstream of compressed image data having the first structure of the group of frames of images, get their first motion vector from an input bitstream of compressed image data to form an output bit stream of the compressed image data based on input image data, wherein receiving the second motion vector from the input bitstream of compressed image data, apply weighting factors to the first and the second motion vector, using the mentioned first motion vector for the extrapolation of the third motion vector for the specified output bit stream of the compressed image data if the weight factor mentioned first motion vector more than the weight of the mentioned second motion vector, when the output bit stream of the compressed image data has a second structure of the group image frames other than the first patterns of the groups of frames of images.

2. The method according to claim 1, characterized in that it further manage what korostil transmission bits of the output bit stream of the compressed image data thus to the bit rate of the input bit stream is different from the bit rate of the output bit stream.

3. The method according to claim 1, characterized in that it further regulate the size of image frames that represent the input bit stream of the compressed image data so that image frames that represent the output bit stream of the compressed image data have a size different from the size of image frames that represent the input bit stream of the compressed image data.

4. The method according to claim 2, characterized in that it further regulate the size of image frames that represent the input bit stream of the compressed image data to frames of images represented the output bit stream of the compressed image data have a size different from the size of image frames that represent the input bit stream.

5. The method according to claim 4, characterized in that it further encode frames of images represented the output bit stream of the compressed image data, as the fields of image frames when images represent the input bit stream of the compressed image data, encoded as full frames of images.

6. The method according to claim 4, characterized in that it further encode frames of images represented output bit thread the compressed image data, as full frames of image frames when images represent the input bit stream of the compressed image data, encoded as fields of the image.

7. The method according to claim 4, characterized in that it further alternate frames of images that represent the input bit stream of the compressed image data, when image frames that represent the incoming bit stream of the compressed image data taken in direct sequence, so that frames of images represented the output bit stream of the compressed image data output in alternating sequence.

8. The method according to claim 4, characterized in that it further relieve the succession of images that represent the input bit stream of the compressed image data, when image frames that represent the input bit stream of compressed data of the image taken in alternating sequence, so that frames of images represented the output bit stream of the compressed image data output in direct sequence.

9. The method according to claim 1, characterized in that it further encode frames of images represented the output bit stream of the compressed image data, as the fields of image frames when images represent the input bit stream of the compressed image data, Cody is ofany as full frames of images.

10. The method according to claim 1, characterized in that it further encode frames of images represented the output bit stream of the compressed image frames as full frames of image frames when images represent the input bit stream of the compressed image frames, encoded as fields of the image.

11. The method according to claim 1, characterized in that it further alternate frames of images that represent the input bit stream of the compressed image data, when image frames that represent the input bit stream of compressed data of the images taken in direct sequence, so that frames of images represented the output bit stream of the compressed image data output in alternating sequence.

12. The method according to claim 1, characterized in that additionally have a succession of images that represent the input bit stream of the compressed image data, when image frames that represent the input bit stream of compressed data of the image taken in alternating sequence, so that frames of images represented the output bit stream of the compressed image data output in direct sequence.

13. The method according to claim 1, characterized in that the specified input bit stream of compressed data of the image is the main image is laid, when the output bit stream of the compressed image data is part of the main image.

14. The method according to claim 1, characterized in that when the first and second motion vectors to determine whether the verified motion vector in a predetermined time range, and get a proven vector movement, if determines that the verified motion vector is within a predetermined time range.

15. The method according to claim 1, characterized in that the weighing of the first and second motion vector is calculated, at least one of the following options: weight of overlap, the time weight, and rounding to the selected motion vector, and receive a weighted value for the selected motion vector based on at least one of the calculated weights.

16. The method of encoding image data, according to which receive the input bitstream of compressed image data that has an input parameter encoding, get the first motion vector from an input bitstream of compressed image data to form an output bit stream of the compressed image data based on input image data, wherein receiving the second motion vector from the input bitstream of compressed image data, apply weighting factors to the first and Oromo the motion vector, use the mentioned first motion vector for the extrapolation of the third motion vector for the specified output bit stream of the compressed image data if the weight factor mentioned first motion vector is greater than the weight of the second motion vector, when the output bit stream is output parameter encoding that differs from the input parameter encoding the input bit stream.

17. The method according to item 16, wherein when the first and second motion vectors to determine whether the verified motion vector in a predetermined time range, and get a proven vector movement, if determines that the verified motion vector is within a predetermined time range.

18. The method according to item 16, characterized in that the weighing of the first and second motion vectors is calculated, at least one of the following parameters: weight of overlap, the time weight, and rounding to the selected motion vector, and receive a weighted value for the selected motion vector based on at least one of the calculated weights.

19. The method according to item 16, characterized in that the weighing of the first and second motion vectors is calculated, at least one of the following parameters: weight of overlap, the time the EU and the weight off for the selected motion vector, and receive a weighted value for the x - and y-components of the selected motion vector based on at least one of the calculated weights.

20. The method according to item 16, characterized in that the input parameter encoding represents at least one of the parameters: the structure of the group image frames, the frame size of the image, bit rate, format, full frame images, the format of the frame images, a direct sequence and alternating the sequence.

21. The method according to item 16, characterized in that the output of the encoding represents at least one of the parameters: the structure of the group image frames, the frame size of the image, bit rate, format, full frame images, the format of the frame images, a direct sequence and alternating the sequence.

22. The method according to item 16, wherein the first and second motion vectors represent the first and second clusters of motion vectors.



 

Same patents:

The invention relates to television technology, in particular to a device for the determination of mobile and immobile blocks of the image between adjacent frames of the image

The invention relates to a method and apparatus for evaluation (determination) movement, and in particular to a method and apparatus for a detailed assessment of traffic that use the deviation generated by the motion estimation to determine the motion vector, especially when using the method of differential pulse code modulation (DICM) for encoding digital image signal

FIELD: coding elementary-wave data by means of null tree.

SUBSTANCE: proposed method includes generation of elementary-wave ratios pointing to image. In the process bits of each elementary-wave ratio are associated with different bit places so that each place is associated with one of bits of each elementary-wave ratio and associated bits are coded with respect to each place of bits to point to null tree roots. Each place of bits is also associated only with one of bits of each elementary-wave ratio. Computer system 100 for coding elementary-wave ratios by means of null tree has processor 112 and memory 118 saving program that enables processor 112 to generate elementary-wave ratios pointing to image. Processor 112 functions to code bits of each place to point to null tree roots associated with place of bits.

EFFECT: enhanced data compression speed.

18 cl, 7 dwg

The invention relates to the field of shorthand, more specifically to the introduction and detection of watermark data in the data stream, such as, for example, video stream

The invention relates to the field of steganography, in particular for detecting a digital watermark

The invention relates to image compression to reduce bandwidth requirements of digital video decoder

Camcorder // 2191483
The invention relates to video equipment, in particular, to the design professional camcorder related to the type of camcorders, the peculiarity of which is the use of image compression and digital recording in the memory the received audio and video data and storing them in memory in a compressed form

The invention relates to encoding and decoding images

FIELD: coding elementary-wave data by means of null tree.

SUBSTANCE: proposed method includes generation of elementary-wave ratios pointing to image. In the process bits of each elementary-wave ratio are associated with different bit places so that each place is associated with one of bits of each elementary-wave ratio and associated bits are coded with respect to each place of bits to point to null tree roots. Each place of bits is also associated only with one of bits of each elementary-wave ratio. Computer system 100 for coding elementary-wave ratios by means of null tree has processor 112 and memory 118 saving program that enables processor 112 to generate elementary-wave ratios pointing to image. Processor 112 functions to code bits of each place to point to null tree roots associated with place of bits.

EFFECT: enhanced data compression speed.

18 cl, 7 dwg

FIELD: converting code of received video sequence using extrapolated movement data received from video sequence.

SUBSTANCE: proposed method for code conversion involves reception of first bit stream of compressed picture data having some coding parameters. These parameters may relate to GF-structure of picture frames, picture frame size, to parameter showing if frames presented in input bit stream are picture fields or frames, and/or if they form picture frames presented in bit stream, direct or interlaced sequence. First and second movement vectors are obtained from input bit stream and used together with weighting coefficients to extrapolate third movement vector for output bit stream of compressed picture data. Output bit stream that differs from input one by one or more parameters is outputted as converted-code output signal.

EFFECT: provision for minimizing of or dispensing with movement estimation in code conversion process.

22 cl, 4 dwg, 1 tbl

FIELD: protection of video information against unauthorized copying.

SUBSTANCE: proposed method using watermarks to protect video information against unauthorized copying by changing scale of pattern in the course of copying includes introduction of watermark in original video signal with different scales. Watermark is maintained in each scale for preset time interval sufficient to enable detector circuit in digital-format video recorder to detect, extract, and process information contained in watermark. Watermark scale is changed by end of preset time interval preferably on pseudorandom basis to ensure appearance of each of all scales in predetermined scale variation range as many times as determined in advance. In this way definite scale possessing ability of watermark recovery to initial position and size can be identified and used for watermark detection.

EFFECT: enhanced reliability, facilitated procedure.

24 cl, 7 dwg

FIELD: multimedia technologies.

SUBSTANCE: method includes at least following stages: determining, whether current value of processed coefficient of discontinuous cosine transformation is equal or less than appropriate threshold value, used in current time for quantizing coefficients of discontinuous cosine transformation of image blocks of common intermediate format, and if that is so, then value of discontinuous cosine transformation coefficient is set equal to zero, then currently used threshold value is increased for use as threshold value with next processing of discontinuous cosine transformation coefficient, in opposite case currently used threshold value is restored to given original threshold value, which is used as threshold value for next processing of discontinuous cosine transformation coefficient; determining, whether increased threshold value is greater than given upper limit of threshold value, and if that is so, than increased threshold value is replaced with given upper limit.

EFFECT: higher quality.

8 cl, 4 dwg

FIELD: methods and devices for memorization and processing of information containing video images following one another.

SUBSTANCE: from each image recorded prior to current time appropriately at least one image area is selected and aperture video information is recorded with placement information. from video-information at least one mixed image is generated with consideration of appropriate placement information. Mixed image is utilized for display in accordance to movement estimation, movement compensation or error masking technology frames.

EFFECT: decreased memory resource requirements for memorization of multiple previously received images.

3 cl, 4 dwg

FIELD: engineering of devices for transforming packet stream of information signals.

SUBSTANCE: information signals represent information, positioned in separate, serial packets of digital format data. These are transformed to stream of information signals with time stamps. After setting of time stamps, which are related to time of arrival of data packet, time stamps of several data packets are grouped as packet of time stamps, wherein, in accordance to realization variant, size of time stamps packet equals size of data block.

EFFECT: improved addition of data about time stamps to data packets with fixed size.

6 cl, 29 dwg

FIELD: engineering of circuit for compressing image signals, using blocks and sub-blocks of adaptively determined sizes of given coefficients of discontinuous cosine transformation.

SUBSTANCE: block size-setting element in encoder selects a block or sub-block of processed input pixels block. Selection is based on dispersion of pixel values. Blocks with dispersions greater than threshold are divided, while blocks with dispersions lesser then threshold are not divided. Transformer element transforms pixel values of selected blocks to frequency range. Values in frequency range may then be quantized, transformed to serial form and encoded with alternating length during preparation for transmission.

EFFECT: improved computing efficiency of image signals compression stages without loss of video signals quality levels.

4 cl, 5 dwg

FIELD: technology for compacting and unpacking video data.

SUBSTANCE: for each pixel of matrix priority value is determined, pixels difference value is calculated, priority values utilized for calculating value of pixels priority are combined in one pixels group, pixel groups are sorted, pixel groups are saved and/or transferred in accordance to their priority in priority matrix, while aforementioned operations are constantly repeated, while values of pixel group priorities are repeatedly determined anew, priority matrix for any given time contains pixel groups sorted by current priorities, and also preferably stored first and transferred are pixel groups, which have highest priority and still were not transferred.

EFFECT: simple and flexible synchronization at different transfer speeds, width of transfer band, resolution capacity and display size, respectively.

2 cl, 8 dwg, 1 tbl

FIELD: engineering of systems for encoding moving images, namely, methods for encoding moving images, directed at increasing efficiency of encoding with use of time-wise remote supporting frames.

SUBSTANCE: method includes receiving index of supporting frame, standing for supporting frame, pointed at by other block, providing movement vector for determining movement vector of current block, and determining movement vector of current block with utilization of supporting frame index, denoting a supporting frame.

EFFECT: increased efficiency of encoding in direct prediction mode, decreased number of information bits for frame, in which scene change occurs.

3 cl, 6 dwg

FIELD: engineering of systems for encoding moving image, namely - methods for encoding moving image, directed at increase of encoding efficiency with use of time-wise remote supporting frames.

SUBSTANCE: in the method in process of encoding/decoding of each block of B-frame in direct prediction mode movement vectors are determined, using movement vector of shifted block in given frame, utilized for encoding/decoding B-frame, and, if type of given frame is time-wise remote supporting frame, one of movement vectors, subject to determining, is taken equal to movement vector of shifted block, while another one of movement vectors, subject to determining, is taken equal to 0.

EFFECT: increased encoding efficiency in direct prediction mode, decreased amount of information bits for frame, wherein a change of scene occurs.

2 cl, 6 dwg

Up!