US20070153904A1 - Motion vector detection method and apparatus - Google Patents

Motion vector detection method and apparatus Download PDF

Info

Publication number
US20070153904A1
US20070153904A1 US11/686,479 US68647907A US2007153904A1 US 20070153904 A1 US20070153904 A1 US 20070153904A1 US 68647907 A US68647907 A US 68647907A US 2007153904 A1 US2007153904 A1 US 2007153904A1
Authority
US
United States
Prior art keywords
block
motion vector
frame
image
blocks
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/686,479
Inventor
Goh Itoh
Nao Mishima
Haruhiko Okumura
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US11/686,479 priority Critical patent/US20070153904A1/en
Publication of US20070153904A1 publication Critical patent/US20070153904A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/144Movement detection
    • H04N5/145Movement estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/521Processing of motion vectors for estimating the reliability of the determined motion vectors or motion vector field, e.g. for smoothing the motion vector field or for correcting motion vectors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/533Motion estimation using multistep search, e.g. 2D-log search or one-at-a-time search [OTS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/573Motion compensation with multiple frame prediction using two or more reference frames in a given prediction direction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures

Definitions

  • the present invention relates to a motion vector detection method and apparatus.
  • an image display unit is classified roughly into an impulse type display that continues to emit light during afterglow time of fluorescent material after writing of image, and a hold model display that continues to display a previous frame till writing of image is performed afresh.
  • an impulse type display that continues to emit light during afterglow time of fluorescent material after writing of image
  • a hold model display that continues to display a previous frame till writing of image is performed afresh.
  • CTR display and a field emission type display (referred to as FED) in the impulse type display.
  • FED field emission type display
  • LCD liquid crystal display unit
  • ELD electroluminescence display
  • problems of the hold type display are indistinct phenomena occurring in an image display.
  • the occurrence of the indistinct phenomenon causes in that the images of a plurality of frames are piled up and projected on the retina when a moving object exists in the images over a plurality of frames, and the eyes of an inspection person followed the movement of the moving object.
  • a display image of a frame continues to be displayed until the display image is changed to that of a frame subsequent thereto.
  • an operator is to observe the display image of the frame while predicting that of the frame subsequent thereto and moving the former display image in a moving direction of the moving object.
  • the tracking motion of eyes of the operator has continuity, and the eyes allows sampling whose interval is shorter than the frame interval. The eyes are therefore casted on an image as if the image were formed between two adjacent frames. As a result, the image is observed as an indistinct image.
  • the shortening of the frame interval is preferable in order to solve the problem. It is conceivable as a concrete technique to perform interpolation between adjacent frames by forming an interpolation image using motion compensation used in MPEG (Motion Picture Experts Group) 1 and MPEG2. In the motion compensation, a motion vector detected by block matching is used.
  • MPEG 1 and MPEG2 premise basically to increase a compression ratio, and whether the motion vector reproduces real movement precisely is not matter. Because the error of a motion vector appears as the error of a prediction signal generated by the motion compensation, and the error signal between the prediction signal and the input image signal is encoded by DCT. However, in an image display system such as a hold type display, when an interpolation image is formed by the motion compensation using the motion vector that is not a precision, the display image is deteriorated.
  • Mr. Tei et al. (Nippon Hoso Kyokai) provides a method of utilizing a data group of rhombuses formed by a plurality of frames and blocks in Jpn. Pat. Appln. KOKAI Publication 10-304369.
  • This technique uses a principle that a usual image performs a uniform motion (including a stationary state) with a plurality of frames, and can expect an improvement of motion vector detection precision.
  • a motion vector detection method of detecting a motion vector between a m-th frame (m is an integer) and a (m+n)-th frame (n is an integer not less than i+1 and i is an integer not less than 1) comprising: computing a plurality of motion vector candidates between the m-th frame and the (m+n)-th frame; scale-converting the motion vector candidates according to a ratio of a frame difference between a (m+i)-th frame and the (m+n)-th frame to a frame difference between the (m+i)-th frame and the m-th frame; forming virtually an image on the (m+i)-th frame according to each of the scale-converted motion vector candidates; and selecting, from the motion vector candidates, an optimum motion vector that maximizes a correlation between the image virtually formed and an actual image on the (m+i)-th frame.
  • a motion vector detection method of detecting a motion vector between a m-th frame (m is an integer) and a (m+n)-th frame (n is an integer not less than i+1 and i is an integer not less than 1) comprising: extracting a first block from a motion vector search area of the m-th frame; extracting a plurality of second blocks having a large correlation with respect to the first block from a motion vector search area of the (m+n)-th frame; detecting a plurality of first motion vectors between the first block and the plurality of second blocks; extracting, from a (m+i)-th frame, a third block that is located in spatially the same position as that of the first block; computing a plurality of second motion vectors that are (n ⁇ 1)/n times the first motion vectors; extracting, from the (m+n)-th frame, a fourth block corresponding to a movement position of the third block according to each of the second motion vector
  • FIG. 1 shows a diagram for explaining a motion vector detection method related to the first embodiment of the present invention
  • FIG. 2 is a flowchart that shows a motion vector detection procedure related to the embodiment
  • FIG. 3 is a diagram illustrating an output result and effect of an image formed by the embodiment
  • FIG. 4 is a diagram illustrating an example of an image from which an error motion vector is detected
  • FIG. 5 is a block diagram of a motion vector detection apparatus related to the embodiment
  • FIG. 6 is a diagram for explaining a motion vector detection method related to the second embodiment of the present invention.
  • FIG. 7 is a flowchart showing a motion vector detection procedure related to the embodiment.
  • FIG. 8 is a block diagram of a motion vector detection apparatus related to the embodiment.
  • FIG. 9 is a diagram for explaining a motion vector detection method related to the third embodiment of the present invention.
  • FIG. 10 is a flowchart showing a motion vector detection procedure related to the embodiment.
  • FIG. 11 is a block diagram of a motion vector detection apparatus related to the embodiment.
  • FIG. 12 is a diagram for explaining a motion vector detection method related to the fourth embodiment of the present invention.
  • FIG. 13 is a flowchart showing a motion vector detection procedure related to the embodiment
  • FIG. 14 is a flowchart showing the motion vector detection procedure related to the embodiment.
  • FIG. 15 is a block diagram of a motion vector detection apparatus related to the embodiment.
  • FIG. 16 is a diagram for explaining a motion vector detection method related to the fifth embodiment of the present invention.
  • FIG. 17 is a flowchart that shows motion vector detection procedure related to the embodiment.
  • FIG. 18 is a block diagram of a motion vector detection apparatus related to the embodiment.
  • FIG. 19 is a diagram for explaining an interpolation image forming method related to the sixth embodiment of the present invention.
  • FIG. 20 is a flowchart that shows an interpolation image forming procedure related to the embodiment
  • FIG. 21 is a diagram for explaining a motion vector detection method related to the seventh embodiment of the present invention.
  • FIG. 22 is a flowchart that shows a motion vector detection procedure related to the embodiment
  • FIG. 23 is a flowchart that shows a motion vector detection procedure related to the embodiment.
  • FIG. 24 is a block diagram of a display system related to an embodiment of the present invention.
  • an optimum motion vector is detected by a procedure as shown in FIG. 2 .
  • the procedure will be described referring to FIGS. 1 and 2 .
  • a first block 11 is extracted from the motion vector search area of the m-th frame 1 (step S 101 ).
  • a plurality of second blocks 12 having the same block size as the first block 11 in the m-th frame 1 and a large correlation with respect to the first block 11 are extracted from the motion vector search area of the (m+n)-th frame 2 (step S 102 ).
  • a method for inspecting the degree of correlation between the blocks in step S 102 is described as follows. An absolute value difference of data is computed every pixel in the block, for example, to obtain a plurality of absolute value differences. The absolute value differences are added to obtain an absolute value difference sum. It is determined that the correlation is large when the absolute value difference sum is small, and small when absolute value difference sum is large.
  • a plurality of first motion vectors D between the first block 11 and the second blocks 12 are detected as a plurality of motion vector candidates (step S 103 ).
  • the third block 13 that is located in spatially the same position as that of the first block 11 in the m-th frame is extracted from a (m+i)-th frame 3 (step S 104 ).
  • the first motion vector D is converted into a motion vector of the third block 13 in the (m+i)-th frame 3 to the (m+n)-th frame 2 .
  • the second motion vectors ((n ⁇ i)/n)*D which were subjected to scale conversion are obtained (step S 105 ).
  • the frame difference is a difference between the frame numbers of two frames.
  • the fourth block 14 corresponding to a movement position of the third block 13 is extracted according to the second motion vector ((n ⁇ i)/n)*D (step S 106 ).
  • the fourth block 14 is a part of the image which is virtually formed on the (m+i) frame according to the second motion vector ((n ⁇ i)/n)*D.
  • the same block as the first block 11 should exist in the (m+i)-th frame 3 .
  • the position of the block is a position of a dotted line block in FIG. 1 . Therefore, relative position relation between the dotted line block and the third block 13 is equal to relative position relation between the second block 12 and the fourth block 14 .
  • a propriety relative to each of a plurality of first motion vectors D is determined using the third block 13 and the fourth block 14 .
  • a motion vector that maximizes a correlation between the third block 13 and the fourth block 14 is obtained.
  • an optimum motion vector that maximizes the correlation between the third block 13 and the fourth block 14 is selected from a plurality of first motion vectors D (step S 107 ).
  • the correlation between the third block 13 and the fourth block 14 is obtained by calculating a sum of absolute value differences of pixels of both blocks. It is determined that when the absolute value difference sum is small, the correlation is large, and when the absolute value difference sum is large, the correlation is small.
  • the first motion vector D that makes the absolute value difference sum E 1 computed by the following equation (1) minimum is obtained as an optimum motion vector.
  • E ⁇ ⁇ 1 ⁇ ⁇ ⁇ ⁇ ⁇ f ⁇ ( X + D , m + n ) - f ⁇ ( X , m ) ⁇ + ⁇ ⁇ ⁇ f ⁇ ( X + ( n - i ) n ⁇ D , m + n ) - f ⁇ ( X , m + i ) ⁇ ⁇ ( 1 )
  • X indicates a position vector of block
  • f (A, b) a pixel value corresponding to a position (A) of each block and a frame (b).
  • ⁇ and ⁇ are weighting factors. The definition of each of these parameters is applied to each embodiment.
  • the first item of the right side of the equation (1) is an absolute difference sum between the first block 11 and the second block 12
  • the second item is an absolute difference sum between the third block 13 and the fourth block 14 .
  • the weighting factors ⁇ and ⁇ are used for weighting the absolute value difference of each block, and represented as ⁇ to make much of the absolute value difference sum (correlation) between the third block 13 and the fourth block 14 basically. This situation is preferable.
  • FIGS. 3 and 4 The effect obtained by the present invention will be described referring to FIGS. 3 and 4 .
  • two motion vectors A and B are detected as the first motion vector D, when a still image “TEST” is displayed from the m-th frame 1 to the (m+2)-th frame 2 as shown in FIG. 3 .
  • the motion vector A is a motion vector between the first block 11 on the (m)-th frame 1 and the second block 12 A on the (m+2)-th frame 2 .
  • the motion vector B is a motion vector between the first block 11 on (m)-th frame 1 and the second block 12 B different from the second block 12 A on (m+2)-th frame 2 .
  • an image 14 A to be displayed at the position of the third block 13 in the (m+1)-th frame 3 is drawn up according to a motion vector between the m-th frame 1 and the (m+1)-th frame 3 that is obtained by subjecting the motion vector to a scale conversion, it is a part of “ES”.
  • an image 14 B to be displayed at the position of the third block 13 of the (m+1)-th frame 2 is drawn up according to a motion vector between the m-th frame 1 and the (m+1)-th frame, which is obtained by subjecting the motion vector B to scale conversion, it is “T”.
  • the images 14 A and 14 B that are drawn up respectively from the motion vectors A and B are compared with the image of actual (m+1)-th frame 3 .
  • the third block 13 is compared with the fourth blocks 14 A and 14 B.
  • the motion vector B is selected as an optimum motion vector by this comparison.
  • the reason why the motion vector A is detected by mistake is influence of noise as well as that the same characters (this example, T) are included in one still image as shown by the example of “TEST”.
  • T the motion vector A is detected.
  • the probability that the motion vector A is erroneously detected is largely decreased by inspecting a correlation between the third block 13 of (m+1)-th frame 3 and the fourth blocks 14 A and 14 B according to the present embodiment. This effect was confirmed by the experiment performed by the inventors.
  • An input image signal 31 is input to a (m+2)-th frame memory 32 , a (m+1)-th frame memory 33 and a m-th frame memory 34 in turn, Image signals 35 , 36 and 37 of the (m+2)-th frame, (m+1)-th frame and m-th frame are read from the memories 32 , 33 and 34 .
  • the motion vector detection unit 40 receives the image signals 35 and 37 of the (m+2)-th and m-th frames, and performs a process of steps S 101 to S 103 in FIG. 2 , namely extraction of the first block from the video signal 37 of the m-th frame, extraction of the second block from the video signal 35 of the (m+2)-th frame and detection of a plurality of first motion vector 41 between the first and the second blocks.
  • a block extraction unit 43 for motion vector determination extracts a block used for determining a motion vector from the video signals 35 and 36 of the (m+2)-th and (m+1)-th frames.
  • An address 39 of a first block position for example, address of the upper left corner of the block position is output from a first block position designation unit 38 to the motion vector detection unit 40 and the block extraction unit 43 .
  • the motion vector scale conversion unit 42 performs a process of step S 105 in FIG. 2 , namely, computation of a plurality of second motion vectors (D/2) 44 that is 1 ⁇ 2 times the first motion vector 41 (D).
  • the block extraction unit 43 performs a process of steps S 104 to S 106 in FIG. 2 , namely, extraction of the third block 13 from video signal 36 of the (m+1)-th frame and extraction of the fourth block 14 from the video signal 35 of the (m+2)-th frame according to the second motion vector 44 , and outputs video signals 45 of the third and fourth blocks.
  • the video signals 45 of the third and fourth blocks from the block extraction unit 43 are inputs to an optimum motion vector determination unit 46 , and performs a process for determining as an optimum motion vector a motion vector that maximizes a correlation between the third and fourth blocks in a plurality of first motion vectors 41 maximum.
  • the motion vector selector 48 selects one of the first motion vectors 41 as a final optimum motion vector 49 according to determination result 47 of the optimum motion vector determination unit 46 (S 107 ).
  • a motion vector can be precisely obtained from the images of 3 frames, i.e., the m-th, (m+i)-th and (m+n)-th frames.
  • FIGS. 6 and 7 Another procedure for detecting an optimum motion vector between a m-th frame 1 and a (m+n)-th frame 2 will be described as the second embodiment of the present invention referring to FIGS. 6 and 7 .
  • the first block 11 is extracted from the motion vector search area of the m-th frame 1 (step S 201 ).
  • a plurality of second blocks 12 having the same block size as the first block 11 in the m-th frame 1 and a large correlation with respect to the first block 11 is extracted from the motion vector search area of the (m+n)-th frame 2 (step S 202 ).
  • a method for inspecting the degree of correlation between the blocks in step S 102 is described hereinafter.
  • An absolute value difference of pixel data is computed every pixel in the block, for example, to obtain a plurality of absolute value differences.
  • the absolute value differences are added to obtain an absolute value difference sum. It is determined that the correlation is large when the absolute value difference sum is small, and small when absolute value difference sum is large.
  • a plurality of first motion vectors D between the first block 11 and the second blocks 12 are detected as a plurality of motion vector candidates (step S 203 ).
  • the steps S 201 to S 203 correspond to the steps S 101 to S 103
  • a plurality of third motion vectors (i/n)*D scale-converted the first motion vector D to a motion vector of the first block 11 in the m-th frame 1 to a (m+i)-th frame 3 are computed based on a ratio of the frame difference n between the m-th frame 1 and the (m+n)-th frame 2 to the frame difference i between the m-th frame 1 and the (m+i)-th frame 3 (step S 204 ).
  • the fifth block 15 which is a movement location of the first block to the (m+i)-th frame 3 is extracted according to the third motion vector (i/n)*D (step S 205 ).
  • the first motion vector D reproduces actual movement
  • the same block as the first block 11 should exist in the (m+i)-th frame 3 , and thus the block is located at the fifth block. Therefore, the fifth block 15 and the first block 11 become equal. For this reason, the propriety of each of the first motion vectors D is determined using the first block 11 and the fifth block 15 in the present embodiment.
  • a correlation between the first block 11 and the fifth block 15 is inspected. Repeating of this process provides a motion vector that maximizes the correlation between the first block 11 and the fifth block 15 .
  • an optimum motion vector that maximizes the correlation between the first block 11 and the fifth block 15 is selected from the plurality of first motion vectors D (step S 206 ).
  • the correlation between the first block 11 and the fifth block 15 is obtained by computing an absolute value difference sum of pixels of both blocks. If the absolute value difference sum is small, the correlation is large, and if it is large, the correlation is small.
  • one first motion vector D that makes the absolute value difference sum E 2 computed by the following equation (2) minimum is determined as an optimum motion vector.
  • E 2
  • the first term of the right hand side of the equation (2) is an absolute value difference sum between the first block 11 and the second block 12 .
  • the second term is an absolute value difference sum between the first block 11 and the fifth block 15 .
  • the motion vector scale conversion unit 42 computes a third motion vector (D/2) 51 that is 1 ⁇ 2 times the first motion vector 41 (D)(S 204 ).
  • the block extraction unit 43 extracts a block for using in a determination of the first motion vector from the video signals 36 and 37 of the (m+1)-th and m-th frames. More specifically, the block extraction unit 43 extracts the fifth block 15 from the video signal 36 of the (m+1)-th frame according to the third motion vector 51 (S 205 ).
  • the video signal 52 of the fifth block from the block extraction unit 43 is input to the optimum motion vector determination unit 46 .
  • the optimum motion vector determination unit 46 selects a motion vector that maximizes the correlation between the first block 11 and the fifth block 15 from a plurality of first motion vectors 41 and determines it as an optimum motion vector.
  • the motion vector selector 48 selects as a final optimum motion vector 35 one of the plurality of first motion vectors 41 according to determination result 47 from the optimum motion vector determination unit 46 .
  • the present embodiment can provide the same effect as the first embodiment.
  • a motion vector detection method of the third embodiment of the present invention will be described referring to FIGS. 9 and 10 .
  • at first to fifth blocks 11 - 15 are extracted according to the procedure (steps S 301 -S 308 ) similar to that of the first and second embodiments.
  • Some of the first motion vectors that make a large correlation between these blocks 11 , 12 and 15 are selected from the first motion vectors D detected in step S 303 , using the first, second and fifth blocks 11 , 12 and 15 (step S 309 ).
  • the propriety of the first motion vectors D selected in step S 309 is determined using the third and fourth blocks 13 and 14 .
  • One of the first motion vectors D that maximizes a correlation between the third and fourth blocks 13 and 14 maximum is selected as an optimum motion vector (step S 310 ).
  • E ⁇ ⁇ 3 ⁇ ⁇ ⁇ ( ⁇ f ⁇ ( X + D , m + n ) - f ⁇ ( X , m ) ⁇ + ⁇ f ⁇ ( X + i n ⁇ D , m + i ) - f ⁇ ( X , m ) ⁇ ) + ⁇ ⁇ ⁇ f ⁇ ( X + ( n - 1 ) n ⁇ D , m + n ) - f ⁇ ( X , m + i ) ⁇ ⁇ ( 3 )
  • the first item of the right side of the equation (3) is an absolute value difference sum between the first and second blocks 11 and 12
  • the second item is an absolute value difference sum between the first and fifth blocks 11 and 15
  • the third item is an absolute value difference sum between the blocks 13 and 14 .
  • Weighting factors ⁇ and ⁇ are used for weighting the absolute value difference of each block, and represented as ⁇ in order to make much of the absolute value difference sum (correlation) between the third block 13 and the fourth block 14 basically. This situation is preferable.
  • the motion vector detection unit 40 performs extraction of the first block from the video signal 37 of the m-th frame, extraction of the second block from the video signal 35 of a (m+2)-th frame, and detection of a plurality of first motion vectors 41 between the first block and the second block (D).
  • the video signal 53 of the first and second blocks is output to the optimum motion vector determination unit 46 (steps S 301 -S 303 ).
  • the block extraction portion 43 extracts blocks used for a determination of a motion vector from the video signals 35 , 36 and 37 of the (m+2)-th, (m+1)-th and m-th frames. More specifically, the block extraction unit 43 extracts the third block 13 from the video signal 36 of a (m+1)-th frame, the fourth block 14 from video signal 35 of a (m+2)-th frame according to the second motion vector 44 , and the fifth block 15 from the video signal 36 of the (m+1)-th frame according to the third motion vector 51 , and outputs video signals 54 of the third, fourth and fifth blocks (step S 306 and S 308 ).
  • the video signals 54 of the third, fourth and fifth blocks from the block extraction unit 43 are inputs to the optimum motion vector determination unit 46 .
  • the optimum motion vector determination unit 46 a plurality of motion vectors that make a large correlation between the first block 11 and the second and fifth blocks 12 and 15 are selected from the first motion vectors 41 .
  • one motion vector that maximizes a correlation between the third block 13 and the fourth block 14 is determined as an optimum motion vector (S 309 ).
  • the motion vector selector 48 selects an optimum motion vector according to determination result 47 from the optimum motion vector determination unit 46 to output it as a final optimum motion vector 49 (S 310 ).
  • the present embodiment enables to detect a motion vector more precisely with the configuration that combines the first and second embodiments.
  • the fourth embodiment of the present invention will be described in conjunction with FIGS. 12-14 .
  • an interpolation image is made up at the position of a (m+i)-th frame 3 between the m-th frame 1 and (m+n)-th frame 2
  • both of the first motion vector D from the m-th frame 1 to a (m+n)-th frame 2 and the fourth motion vector E from the (m+n)-th frame 2 to the m-th frame 1 are obtained.
  • the detection of the first motion vector D is performed similar to the first embodiment (steps S 401 -S 406 and S 412 ).
  • Detection of the fourth motion vector E is performed as follows.
  • the sixth block 16 that is located in spatially the same position as that of the first block 11 are extracted from the (m+n)-th frame (step S 407 ).
  • the seventh block 17 having a large correlation with respect to the sixth block 16 is extracted from a plurality of blocks of the same block size as the sixth block 16 in a motion vector search area in the m-th frame 1 (step S 408 ).
  • the fourth motion vector E between the sixth block 16 and the seventh block 17 is detected (step S 409 ).
  • the correlation between the blocks is obtained by computing an absolute value difference sum of pixels of both blocks. When the absolute value difference sum is small, the correlation is large, and when it is large, the correlation is small.
  • the fifth motion vector (i/n)*E corresponding to the fourth motion vector E scale-converted by movement quantity of the sixth block 16 from the (m+i)-th frame 3 to the m-th frame 1 is computed from a ratio of a frame difference i between the m-th frame 1 and the (m+i)-th frame 3 to a frame difference n between the (m+n)-th frame 2 and the m-th frame 1 (step S 410 ).
  • the eighth block 18 is extracted according to the fifth motion vector (i/n)*E (step S 411 ).
  • the eight block 18 corresponds to a movement position of the third block 13 to the m-th frame 1 which is in the (m+i)-th frame 3 and at the same position as the first block 11 of the m-th frame 1 .
  • the selection of the first optimum motion vector in step S 412 is performed similar to the first embodiment.
  • the absolute value difference sum between the third block 13 and the fourth block 14 is computed to inspect a correlation, and one motion vector that maximizes the correlation between the third block 13 and the fourth block 14 is selected from the first motion vectors.
  • the fourth optimum motion vector in step S 413 one motion vector that maximizes the correlation between the third block 13 and the eighth block 18 is selected from the fourth motion vectors.
  • the absolute value difference sum E 1 between the third block 13 and the fourth block 14 is compared with the absolute value difference sum E 4 between the third block 13 and the eighth block 18 (step S 414 ), and the motion vector that the absolute value difference sum is small is adopted.
  • E 1 is smaller than E 4
  • the optimum first motion vector detected in step S 412 is selected as the optimum motion vector (step S 415 ).
  • E 4 is smaller than E 1
  • the inverse vector ⁇ E of the optimum fourth motion vector detected in step S 413 is selected as the optimum motion vector (step S 416 ).
  • E ⁇ ⁇ 4 ⁇ ⁇ ⁇ ⁇ ⁇ f ⁇ ( X , m + n ) - f ⁇ ( X + E , m ) ⁇ + ⁇ ⁇ ⁇ f ⁇ ( X + i n ⁇ E , m ) - f ⁇ ( X , m + i ) ⁇ ⁇ ( 4 )
  • weighting factors ⁇ and ⁇ use the same as a value used for obtaining E 1 .
  • a motion vector detection unit 60 a motion vector scale conversion unit 62 , a block extraction unit 63 for motion vector determination, and an optimum motion vector determination unit 66 are newly provided.
  • the motion vector detection unit 40 , motion vector scale conversion unit 42 , block extraction unit 43 , and optimum motion vector determination unit 46 are identical to those of the first embodiment, a process of steps S 401 -S 406 and S 412 in FIG. 13 is performed by these units.
  • the motion vector detection unit 60 performs the process of steps S 407 -S 409 in FIG. 13 .
  • the sixth block 16 is extracted in the motion vector search area from the (m+2)-th frame.
  • the seventh block 17 of the same block size as the sixth block 16 is extracted from the motion vector search area of the m-th frame 1 .
  • the fourth motion vector 61 between the sixth block 16 and the seventh block 17 is detected.
  • the position designation of the sixth block 16 is performed by the first block position designation unit 38 .
  • An address of the first block position for example, address 39 of an upper left corner is output to the motion vector detection unit 60 and the block extraction unit 63 .
  • the motion vector scale conversion unit 62 computes a plurality of fifth motion vectors 64 that are 1 ⁇ 2 times the fourth motion vector 61 (S 410 ).
  • the block extraction unit 63 extracts the eighth block 18 from the video signal 37 of the m-th frame according to the fifth motion vector 64 , and outputs a video signal 65 of the eighth block 18 (S 411 ).
  • the video signal 45 A of the third block from the block extraction unit 43 and the video signal 65 of the eighth block from the block extraction unit 63 are input to the optimum motion vector determination unit 66 .
  • the optimum motion vector determination unit 66 determines as the optimum fourth motion vector a motion vector in the fourth motion vectors 61 that maximizes a correlation between the third block 13 and the eighth block 18 (step S 413 ).
  • the motion vector selection unit 68 selects an optimum motion vector according to determination results 47 and 67 from the optimum motion vector determination units 46 and 66 (step S 415 , S 416 ), and outputs the selected motion vector as a final optimum motion vector 69 .
  • the process for acquiring the optimum fourth motion vector from the m-th frame to the (m+n)-th frame in the first embodiment is combined with the process for acquiring the optimum fourth motion vector from the (m+n)-th frame to the m-th frame that is reversed in time with respect to the above process.
  • the optimum one of the optimum fourth optimum motion vectors provided by these processes is selected as a final optimum motion vector, whereby the motion vector detection precision is improved more.
  • the ninth block 19 is extracted from a motion vector search target of the (m+i)-th frame 3 in order to obtain a motion vector between the m-th frame 1 and the (m+n)-th frame 2 (step S 501 ).
  • a plurality of tenth blocks 20 are extracted from the motion vector search area of the m-th frame 1 (step S 502 ).
  • the sixth motion vector F between the ninth block 19 and the tenth block 20 is detected as a motion vector from the ninth block 19 to the m-th frame 1 (step S 503 ).
  • a motion vector from the ninth block 19 to the (m+n)-th frame 2 is computed the seventh motion vector—((n ⁇ i)/i)*F having a direction opposite to the sixth motion vector F and a size satisfying a ratio of a frame difference between the (m+i)-th frame 3 and the (m+n)-th frame 2 (n ⁇ i) to a frame difference i between the m-th frame 1 and the (m+i)-th frame 3 (step S 504 ).
  • the eleventh block 21 corresponding to a movement location of the ninth block 19 is extracted from the (m+n)-th frame 2 according to the seventh motion vector—((n ⁇ i)/i)*F (step S 505 ).
  • a plurality of sixth motion vectors F that make a large correlation between the tenth block 20 and the eleventh block 21 are detected (step S 506 ). If the sixth motion vector F reproduces actual movement, the same block as the tenth block 20 or the eleventh block 21 must exist in (m+i)-th frame 3 . The position of that block is located at the ninth block 19 . In the present embodiment, the image of the ninth block 19 is extracted from the (m+i)-th frame 3 , the propriety of the sixth motion vector F is determined using the images of the ninth block 19 , the tenth block 20 and the eleventh block 21 .
  • the correlation between the tenth block 20 and the eleventh block 21 is inspected, and a plurality of sixth blocks that makes the correlation larger are extracted. Furthermore, the correlation between the ninth block 19 and the tenth block 20 is inspected, and one of the sixth motion vectors that maximize the correlation is selected as the optimum motion vector.
  • E ⁇ ⁇ 5 ⁇ ⁇ ⁇ ⁇ ⁇ f ⁇ ( X + F , m ) - f ⁇ ( X - ( n - i ) i ⁇ F , m + n ) ⁇ + ⁇ ⁇ ⁇ f ⁇ ( X + F , m ) - f ⁇ ( X , m + i ) ⁇ ⁇ ( 5 )
  • the correlation between the ninth block 19 and the tenth block 20 in an equation (5) is inspected in this embodiment, but the correlation between the ninth block 19 and the eleventh block 21 may be inspected.
  • the absolute value difference sum E 6 is expressed by the following equation (6).
  • E ⁇ ⁇ 6 ⁇ ⁇ ⁇ ⁇ ⁇ f ⁇ ( X + F , m ) - f ⁇ ( X - ( n - i ) i ⁇ F , m + n ) ⁇ + ⁇ ⁇ ⁇ f ⁇ ( X - ( n - i ) i ⁇ F , m + n ) - f ⁇ ( X , m + i ) ⁇ ⁇ ( 6 )
  • Weighting factors ⁇ and ⁇ in the equations (5) and (6) are ⁇ . This situation is preferable.
  • E ⁇ ⁇ 7 ⁇ ⁇ ⁇ ⁇ ⁇ f ⁇ ( X + F , m ) - f ⁇ ( X - ( n - i ) i ⁇ F , m + n ) ⁇ + ⁇ ⁇ ⁇ f ⁇ ( X + F , m ) - f ⁇ ( X , m + i ) ⁇ + ⁇ ⁇ ⁇ f ⁇ ( X - ( n - i ) i ⁇ F , m + n ) - f ⁇ ( X , m + i ) ⁇ ⁇ ( 7 )
  • the block having a large correlation in a time direction is attached great importance with respect to the weighting factors ⁇ and ⁇ .
  • n ⁇ i and i are compared to each other, the block between frames corresponding to small one of n ⁇ i and i is attached great importance.
  • a motion vector F that E 5 , E 6 or E 7 of equation (5), (6) and (7) is minimum is obtained as an optimum motion vector.
  • the block extraction unit 83 for motion vector determination extracts the ninth block 19 in a motion vector search area from the video signal 36 of the (m+1)-th frame (S 501 ).
  • the ninth block position designation unit 78 designates the position of the ninth block 19 .
  • An address of the ninth block position, for example, address 79 of an upper left corner is output to the block extraction unit 83 .
  • the motion vector detection unit 80 performs a process of steps S 502 -S 506 .
  • the sixth motion vector F that is a motion vector from the (m+i)-th frame 3 to the m-th frame 1 is detected.
  • the seventh motion vector that is reverse with respect to the motion vector F and is 1 ⁇ 2 thereof is computed.
  • the tenth block 20 of the m-th frame 1 and the eleventh block 21 of the (m+n)-th frame 2 are extracted with respect to the ninth block 19 using the sixth motion vector and the seventh motion vector.
  • the motion vector detection unit 80 outputs a plurality of sixth motion vectors 81 that make a large correlation between the tenth block 20 and the eleventh block 21 and a video signal 82 of the tenth block 20 .
  • the optimum motion vector determination unit 86 determines as an optimum motion vector a motion vector in the sixth motion vectors 81 that maximize the correlation between the ninth block 19 and the tenth block 20 .
  • the motion vector selection unit 88 selects an optimum motion vector according to determination result 87 from the optimum motion vector determination portion 86 (S 507 ) and outputs a selected motion vector as a final optimum motion vector 89 .
  • the configuration that determines the correlation between the tenth block 20 and the ninth block 19 is described in the above embodiment.
  • the correlation between the eleventh block 21 and the ninth block 19 is determined as described above, the determination can be executed by the similar configuration.
  • the interpolation image forming method related to the present embodiment forms an interpolation image by extracting the twelfth block 22 using a motion vector obtained by scale-converting the optimum motion vector D acquired by the first embodiment, and assigning the twelfth block 22 to spatially the same position as the first block of an interpolation frame.
  • an interpolation image is formed at position of a (m+k)-th (k is arbitrary actual number) frame between an m-th frame (m is arbitrary integer) and a (m+n)-th frame (n is an integer not less than i+1 and i is an integer not less than 1) of an original picture, for example, at the position of the (m+k)-th frame between the m-th frame 1 and a (m+i)-th frame 3 as shown in FIG.
  • an optimum motion vector D between the m-th frame 1 and the (m+n)-th frame 2 is acquired, and then the optimum motion vector D is subjected to scale conversion according to a time position of the (m+k)-th frame (steps S 601 and S 602 ) similar to the first embodiment.
  • a movement quantity of the block on the (m+k)-th frame that is located in spatially the same position as that of the first block 11 and moves from the (m+k)-th frame to the (m+i)-th frame 3 that is, a motion vector ((i ⁇ k)/n)*D is computed from a ratio of a frame difference n between the m-th frame 1 and the (m+n)-th frame 2 to a frame difference (i ⁇ k) between the (m+k)-th frame and (m+i)-th frame 3 .
  • the twelfth block 22 corresponding to a movement position of the first block 11 from the (m+i)-th frame 3 that is, twelfth block 22 corresponding to the movement position of the block 13 on the (m+k)-th frame that is located in spatially the same position as that of the first block 11 and moves to the (m+i)-th frame 3 is extracted according to the motion vector ((i ⁇ k)/n)*D after scale conversion (step S 603 ).
  • the twelfth block 22 is assigned to spatially the same position as the first block of the (m+k)-th frame to form an interpolation image of the (m+k)-th frame (step S 604 ). Since the blocks on the (m+k)-th frame are formed of blocks located at spatially the same position from the m-th frame 1 , they are in a manner spread on the (m+k)-th frame. Interpolation images can be formed without clearance and overlapping by forming each interpolation image in correspondence with each block.
  • the interpolation image can be formed on a frame between the m-th frame 1 and the (m+i)-th frame 3 as described before, and also on a frame between the (m+i)-th frame 3 and the (m+n)-th frame 2 using the same motion vector D.
  • the following method is available. At first the (m+0.5)-th frame image and (m+1.5)-th frame image are computed using the motion vector D 1 detected by the m-th frame 1 and (m+2)-th frame, and then the (m+1.5)-th frame image and (m+2.5)-th frame image are computed using another motion vector D 2 detected by the (m+1)-th frame and (m+3)-th frame. Furthermore, a correlation between the (m+1.5)-th frame image and the (m+2)-th frame image is inspected, and each block is selected so that a correlation with respect to the (m+2)-th frame image becomes large.
  • FIGS. 21-23 A method of detecting a motion vector and forming an interpolation image when an original picture is an interlaced picture will be described referring to FIGS. 21-23 .
  • the first block 11 is extracted from the motion vector search area of the m-th field 1 (step S 701 ).
  • a plurality of second blocks 12 having a large correlation with respect to the first block are extracted from a plurality of blocks of the same block size as the first block 11 in the motion vector search area of the (m+2n)-th field (n is an integer more than 1) (step S 702 ).
  • the first vector connecting between the first block 11 and the second block 12 is detected as the motion vector D (step S 703 ).
  • the third block (interlaced block) 13 at a spatial position shifted spatially by 1 ⁇ 2 line in a vertical direction (upper or lower direction) from the first block 11 of the m-th frame 1 is extracted from a (m+2n ⁇ 1)-th field (step S 704 ).
  • a second motion vector D/2 whose direction is the same as that of the first motion vector D and whose quantity of movement is 1 ⁇ 2 that of the first motion vector D (step S 705 ) is computed.
  • the fourth block (interlaced block) 14 at a movement position of a block that is located in spatially the same position as that of the first block 11 and moves to the (m+2n)-th field is extracted from the (m+2n)-th field according to the second motion vector D/2 (step S 706 ).
  • a thirteenth block 23 is formed by subjecting the third block 13 to linear interpolation in the first vertical direction (e.g., upper direction)(step S 707 ). Furthermore, a 14th block 24 is formed by subjecting the fourth block 14 to linear interpolation in the second vertical direction (lower direction) opposite to the first vertical direction (step S 708 ). Propriety of the first motion vector 11 is determined using the 13th block 23 and the 14th block 24 . In other words, a correlation between, for example, the 13th block 23 and the 14th block 24 is inspected by computing an absolute value difference sum, and the first motion vector D that maximizes the correlation is selected as an optimum motion vector (step S 709 ).
  • An interpolation image of the (m+k)-th field is formed using basically the same method as the sixth embodiment.
  • the third motion vector is obtained by scale-converting the optimum motion vector pursued with step S 709 according to a time position of the (m+k)-th field (step S 710 ).
  • the 15th block corresponding to a movement position of the first block 11 from the (m+2n)-th field is extracted from the (m+2n)-th field according to the third motion vector after scale conversion (step S 711 ).
  • the interpolation image of the (m+k)-th field is formed by assigning the 15th block to spatially the same position as the first block 11 of the (m+k)-th field (step S 712 ).
  • the 15th block is an image on the frame (m+n)-th frame. However, when no image exists at the position indicated by the motion vector, that is, an image exists between lines, an image is formed using upper and lower lines.
  • the 15th block is formed by subjecting the upper and lower lines to, for example, a linear interpolation.
  • the fifth block is assigned to the (m+1.5)-th Field.
  • the interpolation image is completed by acquiring the fifth blocks on (m+1.5)-th field and assigning them to the (m+1.5)-th field.
  • This method acquires a progressive image as a field image basically. When the field of interpolation image is subjected to an interlacing process, the field must be vertically skipped with respect to lines.
  • FIG. 24 shows a schematic configuration of an image display system.
  • An input image signal 101 is input to an interpolation field image forming unit 102 and an image switching unit 104 .
  • the interpolation field image forming unit 102 forms an interpolation image signal according to the procedure described above.
  • the interpolation image signal 103 is output to the image switching unit 104 .
  • the image switching unit 104 switches from the input image signal 101 to the interpolation video signal 104 or vice versa.
  • the output video signal 105 from the image switching unit 104 is output to a high-speed refresh display 106 that is a hold type display.
  • the display 106 changes a refresh rate according to a synchronizing signal included in the output video signal 105 , and displays an image.
  • the error motion vector different from actual movement of the image is removed and the precision motion vector can be detected. Further, an output image signal of high frame frequency can be formed from an input image signal of low frame frequency by forming an interpolation image using the precision motion vector. A moving image that is more real and a still picture of a little picture quality degradation can be provided.

Abstract

A motion vector detection method includes extracting a first block from the m-th picture, extracting second blocks having a large correlation with respect to the first block from a (m+n)-th picture ((m+n)>m-th), detecting first motion vectors between the first and second blocks, extracting a third block located in spatially the same position as that of the first block from a (m+i)-th picture ((m+n)>(m+i)>m-th), computing second motion vectors of (n−1)/n times the first motion vectors, extracting a fourth block corresponding to a movement position of the third block from the (m+n)-th picture according to the second motion vector, and selecting an optimum motion vector maximizing a correlation between the third and fourth blocks from the first motion vectors.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a divisional of and claims the benefit of priority under 35 USC §120 from U.S. Ser. No. 10/383,631, filed Mar. 10, 2003 and is based upon and claims the benefit of priority under 35 USC §119 from the Japanese Patent Application No. 2002-073207, filed Mar. 15, 2002, the entire contents of which are incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to a motion vector detection method and apparatus.
  • 2. Description of the Related Art
  • Generally, an image display unit is classified roughly into an impulse type display that continues to emit light during afterglow time of fluorescent material after writing of image, and a hold model display that continues to display a previous frame till writing of image is performed afresh. There are a CRT display and a field emission type display (referred to as FED) in the impulse type display. There are a liquid crystal display unit (referred to as LCD) and an electroluminescence display (refer to as ELD) in a hold type display unit.
  • Problems of the hold type display are indistinct phenomena occurring in an image display. The occurrence of the indistinct phenomenon causes in that the images of a plurality of frames are piled up and projected on the retina when a moving object exists in the images over a plurality of frames, and the eyes of an inspection person followed the movement of the moving object.
  • A display image of a frame continues to be displayed until the display image is changed to that of a frame subsequent thereto. However, an operator is to observe the display image of the frame while predicting that of the frame subsequent thereto and moving the former display image in a moving direction of the moving object. In other words, the tracking motion of eyes of the operator has continuity, and the eyes allows sampling whose interval is shorter than the frame interval. The eyes are therefore casted on an image as if the image were formed between two adjacent frames. As a result, the image is observed as an indistinct image.
  • The shortening of the frame interval is preferable in order to solve the problem. It is conceivable as a concrete technique to perform interpolation between adjacent frames by forming an interpolation image using motion compensation used in MPEG (Motion Picture Experts Group) 1 and MPEG2. In the motion compensation, a motion vector detected by block matching is used.
  • MPEG 1 and MPEG2 premise basically to increase a compression ratio, and whether the motion vector reproduces real movement precisely is not matter. Because the error of a motion vector appears as the error of a prediction signal generated by the motion compensation, and the error signal between the prediction signal and the input image signal is encoded by DCT. However, in an image display system such as a hold type display, when an interpolation image is formed by the motion compensation using the motion vector that is not a precision, the display image is deteriorated.
  • On the other hand, detection methods of high precision are thought since it is required in MPEG 4 to reproduce actual movement using the motion vector. For example, Mr. Tei et al. (Nippon Hoso Kyokai) provides a method of utilizing a data group of rhombuses formed by a plurality of frames and blocks in Jpn. Pat. Appln. KOKAI Publication 10-304369. This technique uses a principle that a usual image performs a uniform motion (including a stationary state) with a plurality of frames, and can expect an improvement of motion vector detection precision.
  • In a motion vector detection technique described by Jpn. Pat. Appln. KOKAI Publication 10-304369, a measure for inspecting whether a detected motion vector is right really is not provided. For this reason, when an interpolation image is formed by motion compensation using an error motion vector, the possibility that the display image has deteriorated greatly is still left. Further, a process for a large number of frames is necessary to improve detection precision of a motion vector, resulting in increasing a processing time.
  • BRIEF SUMMARY OF THE INVENTION
  • According to an aspect of the invention, there is provided a motion vector detection method of detecting a motion vector between a m-th frame (m is an integer) and a (m+n)-th frame (n is an integer not less than i+1 and i is an integer not less than 1), the method comprising: computing a plurality of motion vector candidates between the m-th frame and the (m+n)-th frame; scale-converting the motion vector candidates according to a ratio of a frame difference between a (m+i)-th frame and the (m+n)-th frame to a frame difference between the (m+i)-th frame and the m-th frame; forming virtually an image on the (m+i)-th frame according to each of the scale-converted motion vector candidates; and selecting, from the motion vector candidates, an optimum motion vector that maximizes a correlation between the image virtually formed and an actual image on the (m+i)-th frame.
  • According to another aspect of the invention, there is provided a motion vector detection method of detecting a motion vector between a m-th frame (m is an integer) and a (m+n)-th frame (n is an integer not less than i+1 and i is an integer not less than 1), the method comprising: extracting a first block from a motion vector search area of the m-th frame; extracting a plurality of second blocks having a large correlation with respect to the first block from a motion vector search area of the (m+n)-th frame; detecting a plurality of first motion vectors between the first block and the plurality of second blocks; extracting, from a (m+i)-th frame, a third block that is located in spatially the same position as that of the first block; computing a plurality of second motion vectors that are (n−1)/n times the first motion vectors; extracting, from the (m+n)-th frame, a fourth block corresponding to a movement position of the third block according to each of the second motion vectors; and selecting, from the plurality of first motion vectors, an optimum motion vector that maximizes a correlation between the third block and the fourth block.
  • BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWING
  • FIG. 1 shows a diagram for explaining a motion vector detection method related to the first embodiment of the present invention;
  • FIG. 2 is a flowchart that shows a motion vector detection procedure related to the embodiment;
  • FIG. 3 is a diagram illustrating an output result and effect of an image formed by the embodiment;
  • FIG. 4 is a diagram illustrating an example of an image from which an error motion vector is detected;
  • FIG. 5 is a block diagram of a motion vector detection apparatus related to the embodiment;
  • FIG. 6 is a diagram for explaining a motion vector detection method related to the second embodiment of the present invention;
  • FIG. 7 is a flowchart showing a motion vector detection procedure related to the embodiment;
  • FIG. 8 is a block diagram of a motion vector detection apparatus related to the embodiment;
  • FIG. 9 is a diagram for explaining a motion vector detection method related to the third embodiment of the present invention;
  • FIG. 10 is a flowchart showing a motion vector detection procedure related to the embodiment;
  • FIG. 11 is a block diagram of a motion vector detection apparatus related to the embodiment;
  • FIG. 12 is a diagram for explaining a motion vector detection method related to the fourth embodiment of the present invention;
  • FIG. 13 is a flowchart showing a motion vector detection procedure related to the embodiment;
  • FIG. 14 is a flowchart showing the motion vector detection procedure related to the embodiment;
  • FIG. 15 is a block diagram of a motion vector detection apparatus related to the embodiment;
  • FIG. 16 is a diagram for explaining a motion vector detection method related to the fifth embodiment of the present invention;
  • FIG. 17 is a flowchart that shows motion vector detection procedure related to the embodiment;
  • FIG. 18 is a block diagram of a motion vector detection apparatus related to the embodiment;
  • FIG. 19 is a diagram for explaining an interpolation image forming method related to the sixth embodiment of the present invention;
  • FIG. 20 is a flowchart that shows an interpolation image forming procedure related to the embodiment;
  • FIG. 21 is a diagram for explaining a motion vector detection method related to the seventh embodiment of the present invention;
  • FIG. 22 is a flowchart that shows a motion vector detection procedure related to the embodiment;
  • FIG. 23 is a flowchart that shows a motion vector detection procedure related to the embodiment; and
  • FIG. 24 is a block diagram of a display system related to an embodiment of the present invention.
  • DETAILED DESCRIPTION OF THE INVENTION
  • There will now be described an embodiment of the present invention in conjunction with drawings.
  • The First Embodiment
  • As shown in FIG. 1, it is thought to detect an optimum motion vector between an m-th frame (or m-th picture) 1 and a (m+n)-th frame (or (m+n) picture) 2 to form an interpolation image at a time position of a (m+i)-th frame (or (m+i)-th picture) 3 between the m-th frame 1 and the (m+n)-th (n is an integer not less than i+1 and i is an integer not less than 1) frame 2 of an original picture. In the first embodiment of the present invention, an optimum motion vector is detected by a procedure as shown in FIG. 2. The procedure will be described referring to FIGS. 1 and 2.
  • At first, a first block 11 is extracted from the motion vector search area of the m-th frame 1 (step S101). A plurality of second blocks 12 having the same block size as the first block 11 in the m-th frame 1 and a large correlation with respect to the first block 11 are extracted from the motion vector search area of the (m+n)-th frame 2 (step S102). A method for inspecting the degree of correlation between the blocks in step S102 is described as follows. An absolute value difference of data is computed every pixel in the block, for example, to obtain a plurality of absolute value differences. The absolute value differences are added to obtain an absolute value difference sum. It is determined that the correlation is large when the absolute value difference sum is small, and small when absolute value difference sum is large.
  • A plurality of first motion vectors D between the first block 11 and the second blocks 12 are detected as a plurality of motion vector candidates (step S103). The third block 13 that is located in spatially the same position as that of the first block 11 in the m-th frame is extracted from a (m+i)-th frame 3 (step S104). From a ratio of a frame difference between the (m+i)-th frame 3 and the (m+n)-th frame 2 to a frame difference i between the m frame 1 and the (m+i)-th frame 3, the first motion vector D is converted into a motion vector of the third block 13 in the (m+i)-th frame 3 to the (m+n)-th frame 2. In other words, the second motion vectors ((n−i)/n)*D which were subjected to scale conversion are obtained (step S105). The frame difference is a difference between the frame numbers of two frames.
  • The fourth block 14 corresponding to a movement position of the third block 13 is extracted according to the second motion vector ((n−i)/n)*D (step S106). The fourth block 14 is a part of the image which is virtually formed on the (m+i) frame according to the second motion vector ((n−i)/n)*D.
  • If the first motion vector D reproduces actual movement, the same block as the first block 11 should exist in the (m+i)-th frame 3. The position of the block is a position of a dotted line block in FIG. 1. Therefore, relative position relation between the dotted line block and the third block 13 is equal to relative position relation between the second block 12 and the fourth block 14. In the present embodiment, a propriety relative to each of a plurality of first motion vectors D is determined using the third block 13 and the fourth block 14.
  • By repeating a process for inspecting a correlation between the third block 13 and the fourth block 14, a motion vector that maximizes a correlation between the third block 13 and the fourth block 14 is obtained. In other words, an optimum motion vector that maximizes the correlation between the third block 13 and the fourth block 14 is selected from a plurality of first motion vectors D (step S107). The correlation between the third block 13 and the fourth block 14 is obtained by calculating a sum of absolute value differences of pixels of both blocks. It is determined that when the absolute value difference sum is small, the correlation is large, and when the absolute value difference sum is large, the correlation is small.
  • In the motion vector detection method of the present embodiment, the first motion vector D that makes the absolute value difference sum E1 computed by the following equation (1) minimum is obtained as an optimum motion vector. E 1 = { α f ( X + D , m + n ) - f ( X , m ) + β f ( X + ( n - i ) n D , m + n ) - f ( X , m + i ) } ( 1 )
    where X indicates a position vector of block, f (A, b) a pixel value corresponding to a position (A) of each block and a frame (b). α and β are weighting factors. The definition of each of these parameters is applied to each embodiment. The first item of the right side of the equation (1) is an absolute difference sum between the first block 11 and the second block 12, and the second item is an absolute difference sum between the third block 13 and the fourth block 14. The weighting factors α and β are used for weighting the absolute value difference of each block, and represented as β≧α to make much of the absolute value difference sum (correlation) between the third block 13 and the fourth block 14 basically. This situation is preferable.
  • The effect obtained by the present invention will be described referring to FIGS. 3 and 4. Assuming that i=1 and n=2. Further, assuming that two motion vectors A and B are detected as the first motion vector D, when a still image “TEST” is displayed from the m-th frame 1 to the (m+2)-th frame 2 as shown in FIG. 3. In this case, the motion vector A is a motion vector between the first block 11 on the (m)-th frame 1 and the second block 12A on the (m+2)-th frame 2. The motion vector B is a motion vector between the first block 11 on (m)-th frame 1 and the second block 12B different from the second block 12A on (m+2)-th frame 2.
  • If an image 14A to be displayed at the position of the third block 13 in the (m+1)-th frame 3 is drawn up according to a motion vector between the m-th frame 1 and the (m+1)-th frame 3 that is obtained by subjecting the motion vector to a scale conversion, it is a part of “ES”. On the other hand, when an image 14B to be displayed at the position of the third block 13 of the (m+1)-th frame 2 is drawn up according to a motion vector between the m-th frame 1 and the (m+1)-th frame, which is obtained by subjecting the motion vector B to scale conversion, it is “T”.
  • In the present embodiment, the images 14A and 14B that are drawn up respectively from the motion vectors A and B are compared with the image of actual (m+1)-th frame 3. In other words, the third block 13 is compared with the fourth blocks 14A and 14B. The motion vector B is selected as an optimum motion vector by this comparison. The reason why the motion vector A is detected by mistake is influence of noise as well as that the same characters (this example, T) are included in one still image as shown by the example of “TEST”. When noise occurs in “T” to be actually detected in the (m+2)-th frame 2 and a little noise is in “T” indicated by the motion vector as shown in FIG. 4, for example, the motion vector A is detected. In contrast, the probability that the motion vector A is erroneously detected is largely decreased by inspecting a correlation between the third block 13 of (m+1)-th frame 3 and the fourth blocks 14A and 14B according to the present embodiment. This effect was confirmed by the experiment performed by the inventors.
  • A configuration of a motion vector detection apparatus that executes the motion vector detection method related to the present embodiment is shown in FIG. 5. Assuming that i=1, n=2 in order to simplify description here. An input image signal 31 is input to a (m+2)-th frame memory 32, a (m+1)-th frame memory 33 and a m-th frame memory 34 in turn, Image signals 35, 36 and 37 of the (m+2)-th frame, (m+1)-th frame and m-th frame are read from the memories 32, 33 and 34.
  • The motion vector detection unit 40 receives the image signals 35 and 37 of the (m+2)-th and m-th frames, and performs a process of steps S101 to S103 in FIG. 2, namely extraction of the first block from the video signal 37 of the m-th frame, extraction of the second block from the video signal 35 of the (m+2)-th frame and detection of a plurality of first motion vector 41 between the first and the second blocks. A block extraction unit 43 for motion vector determination extracts a block used for determining a motion vector from the video signals 35 and 36 of the (m+2)-th and (m+1)-th frames. An address 39 of a first block position, for example, address of the upper left corner of the block position is output from a first block position designation unit 38 to the motion vector detection unit 40 and the block extraction unit 43.
  • The motion vector scale conversion unit 42 performs a process of step S105 in FIG. 2, namely, computation of a plurality of second motion vectors (D/2) 44 that is ½ times the first motion vector 41 (D). The block extraction unit 43 performs a process of steps S104 to S106 in FIG. 2, namely, extraction of the third block 13 from video signal 36 of the (m+1)-th frame and extraction of the fourth block 14 from the video signal 35 of the (m+2)-th frame according to the second motion vector 44, and outputs video signals 45 of the third and fourth blocks.
  • The video signals 45 of the third and fourth blocks from the block extraction unit 43 are inputs to an optimum motion vector determination unit 46, and performs a process for determining as an optimum motion vector a motion vector that maximizes a correlation between the third and fourth blocks in a plurality of first motion vectors 41 maximum. The motion vector selector 48 selects one of the first motion vectors 41 as a final optimum motion vector 49 according to determination result 47 of the optimum motion vector determination unit 46 (S107).
  • In the present embodiment described above, a motion vector can be precisely obtained from the images of 3 frames, i.e., the m-th, (m+i)-th and (m+n)-th frames. In other words, there is computed a correlation between the third block extracted from the (m+i)-th frame and the fourth block extracted from the (m+n)-th frame using the second motion vector obtained by subjecting a plurality of motion vectors that are candidates of a motion vector to scale conversion, to inspect propriety of each first motion vector. It is possible to detect the motion vector in precision by detecting only a motion vector indicating the maximum correlation as the optimum motion vector and removing motion vectors aside from it.
  • The Second Embodiment
  • Another procedure for detecting an optimum motion vector between a m-th frame 1 and a (m+n)-th frame 2 will be described as the second embodiment of the present invention referring to FIGS. 6 and 7.
  • At first, the first block 11 is extracted from the motion vector search area of the m-th frame 1 (step S201). A plurality of second blocks 12 having the same block size as the first block 11 in the m-th frame 1 and a large correlation with respect to the first block 11 is extracted from the motion vector search area of the (m+n)-th frame 2 (step S202). A method for inspecting the degree of correlation between the blocks in step S102 is described hereinafter. An absolute value difference of pixel data is computed every pixel in the block, for example, to obtain a plurality of absolute value differences. The absolute value differences are added to obtain an absolute value difference sum. It is determined that the correlation is large when the absolute value difference sum is small, and small when absolute value difference sum is large. A plurality of first motion vectors D between the first block 11 and the second blocks 12 are detected as a plurality of motion vector candidates (step S203). The steps S201 to S203 correspond to the steps S101 to S103 of the first embodiment.
  • In the present embodiment, a plurality of third motion vectors (i/n)*D scale-converted the first motion vector D to a motion vector of the first block 11 in the m-th frame 1 to a (m+i)-th frame 3 are computed based on a ratio of the frame difference n between the m-th frame 1 and the (m+n)-th frame 2 to the frame difference i between the m-th frame 1 and the (m+i)-th frame 3 (step S204). The fifth block 15 which is a movement location of the first block to the (m+i)-th frame 3 is extracted according to the third motion vector (i/n)*D (step S205).
  • If the first motion vector D reproduces actual movement, the same block as the first block 11 should exist in the (m+i)-th frame 3, and thus the block is located at the fifth block. Therefore, the fifth block 15 and the first block 11 become equal. For this reason, the propriety of each of the first motion vectors D is determined using the first block 11 and the fifth block 15 in the present embodiment.
  • Concretely, a correlation between the first block 11 and the fifth block 15 is inspected. Repeating of this process provides a motion vector that maximizes the correlation between the first block 11 and the fifth block 15. In other words, an optimum motion vector that maximizes the correlation between the first block 11 and the fifth block 15 is selected from the plurality of first motion vectors D (step S206). The correlation between the first block 11 and the fifth block 15 is obtained by computing an absolute value difference sum of pixels of both blocks. If the absolute value difference sum is small, the correlation is large, and if it is large, the correlation is small.
  • In the motion vector detection method of the present embodiment described above, one first motion vector D that makes the absolute value difference sum E2 computed by the following equation (2) minimum is determined as an optimum motion vector.
    E2=Σ{α|f(X+D,m+n)−f(X,m)|+β|f(X+i/nD,m+i)−f(X,m)|}  (2)
  • The first term of the right hand side of the equation (2) is an absolute value difference sum between the first block 11 and the second block 12. The second term is an absolute value difference sum between the first block 11 and the fifth block 15. When weighting factors α and β are used for weighting the absolute value difference of each block. Fundamentally β≧α is set to make much of the correlation between the first block 11 and the fifth block 15. This situation is preferable.
  • FIG. 8 shows a configuration of a motion vector detection apparatus that executes the motion vector detection method related to the present embodiment. Assuming that i=1 and n=2 in order to simplify description. The difference between the present embodiment and the first embodiment will be described referring to the same references to elements corresponding to elements shown in FIG. 5. The motion vector scale conversion unit 42 computes a third motion vector (D/2) 51 that is ½ times the first motion vector 41 (D)(S204). The block extraction unit 43 extracts a block for using in a determination of the first motion vector from the video signals 36 and 37 of the (m+1)-th and m-th frames. More specifically, the block extraction unit 43 extracts the fifth block 15 from the video signal 36 of the (m+1)-th frame according to the third motion vector 51 (S205).
  • The video signal 52 of the fifth block from the block extraction unit 43 is input to the optimum motion vector determination unit 46. The optimum motion vector determination unit 46 selects a motion vector that maximizes the correlation between the first block 11 and the fifth block 15 from a plurality of first motion vectors 41 and determines it as an optimum motion vector. The motion vector selector 48 selects as a final optimum motion vector 35 one of the plurality of first motion vectors 41 according to determination result 47 from the optimum motion vector determination unit 46. The present embodiment can provide the same effect as the first embodiment.
  • The Third Embodiment
  • A motion vector detection method of the third embodiment of the present invention will be described referring to FIGS. 9 and 10. In the present embodiment, at first to fifth blocks 11-15 are extracted according to the procedure (steps S301-S308) similar to that of the first and second embodiments.
  • Some of the first motion vectors that make a large correlation between these blocks 11, 12 and 15 are selected from the first motion vectors D detected in step S303, using the first, second and fifth blocks 11, 12 and 15 (step S309). The propriety of the first motion vectors D selected in step S309 is determined using the third and fourth blocks 13 and 14. One of the first motion vectors D that maximizes a correlation between the third and fourth blocks 13 and 14 maximum is selected as an optimum motion vector (step S310).
  • Therefore, in the above mentioned motion vector detection method of the present embodiment, one motion vector that an absolute value difference sum E3 computed according to the following equation (3) becomes minimum is obtained as an optimum motion vector. E 3 = { α ( f ( X + D , m + n ) - f ( X , m ) + f ( X + i n D , m + i ) - f ( X , m ) ) + β f ( X + ( n - 1 ) n D , m + n ) - f ( X , m + i ) } ( 3 )
  • The first item of the right side of the equation (3) is an absolute value difference sum between the first and second blocks 11 and 12, the second item is an absolute value difference sum between the first and fifth blocks 11 and 15, and the third item is an absolute value difference sum between the blocks 13 and 14. Weighting factors α and β are used for weighting the absolute value difference of each block, and represented as β≧α in order to make much of the absolute value difference sum (correlation) between the third block 13 and the fourth block 14 basically. This situation is preferable.
  • FIG. 11 shows a configuration of a motion vector detection apparatus which executes a motion vector detection method related to the present embodiment. Assuming that i=1 and n=2 in order to simplify description. The difference between the present embodiment and the first and second embodiments will be described referring to the same references to elements corresponding to elements shown in FIGS. 5 and 8. The motion vector detection unit 40 performs extraction of the first block from the video signal 37 of the m-th frame, extraction of the second block from the video signal 35 of a (m+2)-th frame, and detection of a plurality of first motion vectors 41 between the first block and the second block (D). The video signal 53 of the first and second blocks is output to the optimum motion vector determination unit 46 (steps S301-S303).
  • The motion vector scale conversion unit 42 computes the second motion vector 44 that is (n−i)/n of the first motion vector and the third motion vector 51 that is i/n of the first motion vector. However, since i=1 and n=2 s, a motion vector (D/2) that is ½ times the first motion vector (D) is calculated as the second and third motion vectors (steps S305 and S307).
  • The block extraction portion 43 extracts blocks used for a determination of a motion vector from the video signals 35, 36 and 37 of the (m+2)-th, (m+1)-th and m-th frames. More specifically, the block extraction unit 43 extracts the third block 13 from the video signal 36 of a (m+1)-th frame, the fourth block 14 from video signal 35 of a (m+2)-th frame according to the second motion vector 44, and the fifth block 15 from the video signal 36 of the (m+1)-th frame according to the third motion vector 51, and outputs video signals 54 of the third, fourth and fifth blocks (step S306 and S308).
  • The video signals 54 of the third, fourth and fifth blocks from the block extraction unit 43 are inputs to the optimum motion vector determination unit 46. In the optimum motion vector determination unit 46, a plurality of motion vectors that make a large correlation between the first block 11 and the second and fifth blocks 12 and 15 are selected from the first motion vectors 41. Furthermore, one motion vector that maximizes a correlation between the third block 13 and the fourth block 14 is determined as an optimum motion vector (S309). The motion vector selector 48 selects an optimum motion vector according to determination result 47 from the optimum motion vector determination unit 46 to output it as a final optimum motion vector 49 (S310).
  • As thus described, the present embodiment enables to detect a motion vector more precisely with the configuration that combines the first and second embodiments.
  • The Fourth Embodiment
  • The fourth embodiment of the present invention will be described in conjunction with FIGS. 12-14. In the present embodiment, when an interpolation image is made up at the position of a (m+i)-th frame 3 between the m-th frame 1 and (m+n)-th frame 2, both of the first motion vector D from the m-th frame 1 to a (m+n)-th frame 2 and the fourth motion vector E from the (m+n)-th frame 2 to the m-th frame 1 are obtained. The detection of the first motion vector D is performed similar to the first embodiment (steps S401-S406 and S412).
  • Detection of the fourth motion vector E is performed as follows. The sixth block 16 that is located in spatially the same position as that of the first block 11 are extracted from the (m+n)-th frame (step S407). The seventh block 17 having a large correlation with respect to the sixth block 16 is extracted from a plurality of blocks of the same block size as the sixth block 16 in a motion vector search area in the m-th frame 1 (step S408). The fourth motion vector E between the sixth block 16 and the seventh block 17 is detected (step S409). The correlation between the blocks is obtained by computing an absolute value difference sum of pixels of both blocks. When the absolute value difference sum is small, the correlation is large, and when it is large, the correlation is small.
  • The fifth motion vector (i/n)*E corresponding to the fourth motion vector E scale-converted by movement quantity of the sixth block 16 from the (m+i)-th frame 3 to the m-th frame 1 is computed from a ratio of a frame difference i between the m-th frame 1 and the (m+i)-th frame 3 to a frame difference n between the (m+n)-th frame 2 and the m-th frame 1 (step S410). The eighth block 18 is extracted according to the fifth motion vector (i/n)*E (step S411). The eight block 18 corresponds to a movement position of the third block 13 to the m-th frame 1 which is in the (m+i)-th frame 3 and at the same position as the first block 11 of the m-th frame 1.
  • The selection of the first optimum motion vector in step S412 is performed similar to the first embodiment. In other words, the absolute value difference sum between the third block 13 and the fourth block 14 is computed to inspect a correlation, and one motion vector that maximizes the correlation between the third block 13 and the fourth block 14 is selected from the first motion vectors. In selection of the fourth optimum motion vector in step S413, one motion vector that maximizes the correlation between the third block 13 and the eighth block 18 is selected from the fourth motion vectors.
  • The absolute value difference sum E1 between the third block 13 and the fourth block 14 is compared with the absolute value difference sum E4 between the third block 13 and the eighth block 18 (step S414), and the motion vector that the absolute value difference sum is small is adopted. In other words, when E1 is smaller than E4, the optimum first motion vector detected in step S412 is selected as the optimum motion vector (step S415). When E4 is smaller than E1, the inverse vector −E of the optimum fourth motion vector detected in step S413 is selected as the optimum motion vector (step S416).
  • The absolute value difference sum E4 computed for obtaining the fourth motion vector in the present embodiment can be expressed by the following equation (4). E 4 = { α f ( X , m + n ) - f ( X + E , m ) + β f ( X + i n E , m ) - f ( X , m + i ) } ( 4 )
  • It is desirable that weighting factors α and β use the same as a value used for obtaining E1.
  • FIG. 15 shows a configuration of a motion vector detection apparatus which executes a motion vector detection method related to the present embodiment. Assuming that i=1 and n=2 in order to simplify description. The difference between the present embodiment and the first embodiment will be described referring to the same references to elements corresponding to elements shown in FIG. 5. In FIG. 15, a motion vector detection unit 60, a motion vector scale conversion unit 62, a block extraction unit 63 for motion vector determination, and an optimum motion vector determination unit 66 are newly provided. The motion vector detection unit 40, motion vector scale conversion unit 42, block extraction unit 43, and optimum motion vector determination unit 46 are identical to those of the first embodiment, a process of steps S401-S406 and S412 in FIG. 13 is performed by these units.
  • The motion vector detection unit 60 performs the process of steps S407-S409 in FIG. 13. In other words, the sixth block 16 is extracted in the motion vector search area from the (m+2)-th frame. The seventh block 17 of the same block size as the sixth block 16 is extracted from the motion vector search area of the m-th frame 1. The fourth motion vector 61 between the sixth block 16 and the seventh block 17 is detected.
  • The position designation of the sixth block 16 is performed by the first block position designation unit 38. An address of the first block position, for example, address 39 of an upper left corner is output to the motion vector detection unit 60 and the block extraction unit 63. The motion vector scale conversion unit 62 computes a plurality of fifth motion vectors 64 that are ½ times the fourth motion vector 61 (S410). The block extraction unit 63 extracts the eighth block 18 from the video signal 37 of the m-th frame according to the fifth motion vector 64, and outputs a video signal 65 of the eighth block 18 (S411).
  • The video signal 45A of the third block from the block extraction unit 43 and the video signal 65 of the eighth block from the block extraction unit 63 are input to the optimum motion vector determination unit 66. The optimum motion vector determination unit 66 determines as the optimum fourth motion vector a motion vector in the fourth motion vectors 61 that maximizes a correlation between the third block 13 and the eighth block 18 (step S413).
  • The motion vector selection unit 68 selects an optimum motion vector according to determination results 47 and 67 from the optimum motion vector determination units 46 and 66 (step S415, S416), and outputs the selected motion vector as a final optimum motion vector 69.
  • According to the present embodiment, the process for acquiring the optimum fourth motion vector from the m-th frame to the (m+n)-th frame in the first embodiment is combined with the process for acquiring the optimum fourth motion vector from the (m+n)-th frame to the m-th frame that is reversed in time with respect to the above process. The optimum one of the optimum fourth optimum motion vectors provided by these processes is selected as a final optimum motion vector, whereby the motion vector detection precision is improved more.
  • The Fifth Embodiment
  • The fifth embodiment of the present invention will be described referring to FIGS. 16 and 17. In the present embodiment, at first the ninth block 19 is extracted from a motion vector search target of the (m+i)-th frame 3 in order to obtain a motion vector between the m-th frame 1 and the (m+n)-th frame 2 (step S501). A plurality of tenth blocks 20 are extracted from the motion vector search area of the m-th frame 1 (step S502).
  • The sixth motion vector F between the ninth block 19 and the tenth block 20 is detected as a motion vector from the ninth block 19 to the m-th frame 1 (step S503). As a motion vector from the ninth block 19 to the (m+n)-th frame 2 is computed the seventh motion vector—((n−i)/i)*F having a direction opposite to the sixth motion vector F and a size satisfying a ratio of a frame difference between the (m+i)-th frame 3 and the (m+n)-th frame 2 (n−i) to a frame difference i between the m-th frame 1 and the (m+i)-th frame 3 (step S504). The eleventh block 21 corresponding to a movement location of the ninth block 19 is extracted from the (m+n)-th frame 2 according to the seventh motion vector—((n−i)/i)*F (step S505).
  • A plurality of sixth motion vectors F that make a large correlation between the tenth block 20 and the eleventh block 21 are detected (step S506). If the sixth motion vector F reproduces actual movement, the same block as the tenth block 20 or the eleventh block 21 must exist in (m+i)-th frame 3. The position of that block is located at the ninth block 19. In the present embodiment, the image of the ninth block 19 is extracted from the (m+i)-th frame 3, the propriety of the sixth motion vector F is determined using the images of the ninth block 19, the tenth block 20 and the eleventh block 21. Concretely, the correlation between the tenth block 20 and the eleventh block 21 is inspected, and a plurality of sixth blocks that makes the correlation larger are extracted. Furthermore, the correlation between the ninth block 19 and the tenth block 20 is inspected, and one of the sixth motion vectors that maximize the correlation is selected as the optimum motion vector.
  • The absolute value difference sum E5 that is computed in the present embodiment is expressed by the following equation (5). E 5 = { α f ( X + F , m ) - f ( X - ( n - i ) i F , m + n ) + β f ( X + F , m ) - f ( X , m + i ) } ( 5 )
  • The correlation between the ninth block 19 and the tenth block 20 in an equation (5) is inspected in this embodiment, but the correlation between the ninth block 19 and the eleventh block 21 may be inspected. In this case, the absolute value difference sum E6 is expressed by the following equation (6). E 6 = { α f ( X + F , m ) - f ( X - ( n - i ) i F , m + n ) + β f ( X - ( n - i ) i F , m + n ) - f ( X , m + i ) } ( 6 )
  • Weighting factors α and β in the equations (5) and (6) are β≧α. This situation is preferable.
  • Furthermore, the correlation can be inspected from the ninth and tenth blocks 19 and 20 and the ninth and eleventh blocks 19 and 21. In this case, the absolute value difference sum E7 is expressed by the following equation (7). E 7 = { α f ( X + F , m ) - f ( X - ( n - i ) i F , m + n ) + β f ( X + F , m ) - f ( X , m + i ) + γ f ( X - ( n - i ) i F , m + n ) - f ( X , m + i ) } ( 7 )
  • The block having a large correlation in a time direction is attached great importance with respect to the weighting factors α and β. In other words, n−i and i are compared to each other, the block between frames corresponding to small one of n−i and i is attached great importance. For example, in the event of i=1 and n=3, the correlation between the ninth block 19 and the tenth block 20 is attached great importance because of n−i=2. Therefore, it is preferable that α is larger than β.
  • In the present embodiment, a motion vector F that E5, E6 or E7 of equation (5), (6) and (7) is minimum is obtained as an optimum motion vector.
  • FIG. 18 shows a configuration of a motion vector detection apparatus that executes the motion vector detection method related to the present embodiment. Assuming that i=1 and n=2 in order to simplify description. The difference between the present embodiment and the first and second embodiments will be described referring to the same references to elements corresponding to elements shown in FIG. 5.
  • The block extraction unit 83 for motion vector determination extracts the ninth block 19 in a motion vector search area from the video signal 36 of the (m+1)-th frame (S501). The ninth block position designation unit 78 designates the position of the ninth block 19. An address of the ninth block position, for example, address 79 of an upper left corner is output to the block extraction unit 83.
  • The motion vector detection unit 80 performs a process of steps S502-S506. In other words, the sixth motion vector F that is a motion vector from the (m+i)-th frame 3 to the m-th frame 1 is detected. The seventh motion vector that is reverse with respect to the motion vector F and is ½ thereof is computed. The tenth block 20 of the m-th frame 1 and the eleventh block 21 of the (m+n)-th frame 2 are extracted with respect to the ninth block 19 using the sixth motion vector and the seventh motion vector. The motion vector detection unit 80 outputs a plurality of sixth motion vectors 81 that make a large correlation between the tenth block 20 and the eleventh block 21 and a video signal 82 of the tenth block 20.
  • The optimum motion vector determination unit 86 determines as an optimum motion vector a motion vector in the sixth motion vectors 81 that maximize the correlation between the ninth block 19 and the tenth block 20. The motion vector selection unit 88 selects an optimum motion vector according to determination result 87 from the optimum motion vector determination portion 86 (S507) and outputs a selected motion vector as a final optimum motion vector 89.
  • The configuration that determines the correlation between the tenth block 20 and the ninth block 19 is described in the above embodiment. When the correlation between the eleventh block 21 and the ninth block 19 is determined as described above, the determination can be executed by the similar configuration.
  • The Sixth Embodiment
  • An interpolation image forming method with the use of a motion compensation based on a motion vector detection will be explained as the sixth embodiment of the present invention in conjunction with FIGS. 19 and 20.
  • The interpolation image forming method related to the present embodiment forms an interpolation image by extracting the twelfth block 22 using a motion vector obtained by scale-converting the optimum motion vector D acquired by the first embodiment, and assigning the twelfth block 22 to spatially the same position as the first block of an interpolation frame. More specifically, when an interpolation image is formed at position of a (m+k)-th (k is arbitrary actual number) frame between an m-th frame (m is arbitrary integer) and a (m+n)-th frame (n is an integer not less than i+1 and i is an integer not less than 1) of an original picture, for example, at the position of the (m+k)-th frame between the m-th frame 1 and a (m+i)-th frame 3 as shown in FIG. 19, at first an optimum motion vector D between the m-th frame 1 and the (m+n)-th frame 2 is acquired, and then the optimum motion vector D is subjected to scale conversion according to a time position of the (m+k)-th frame (steps S601 and S602) similar to the first embodiment. In other words, a movement quantity of the block on the (m+k)-th frame that is located in spatially the same position as that of the first block 11 and moves from the (m+k)-th frame to the (m+i)-th frame 3, that is, a motion vector ((i−k)/n)*D is computed from a ratio of a frame difference n between the m-th frame 1 and the (m+n)-th frame 2 to a frame difference (i−k) between the (m+k)-th frame and (m+i)-th frame 3.
  • The twelfth block 22 corresponding to a movement position of the first block 11 from the (m+i)-th frame 3, that is, twelfth block 22 corresponding to the movement position of the block 13 on the (m+k)-th frame that is located in spatially the same position as that of the first block 11 and moves to the (m+i)-th frame 3 is extracted according to the motion vector ((i−k)/n)*D after scale conversion (step S603).
  • The twelfth block 22 is assigned to spatially the same position as the first block of the (m+k)-th frame to form an interpolation image of the (m+k)-th frame (step S604). Since the blocks on the (m+k)-th frame are formed of blocks located at spatially the same position from the m-th frame 1, they are in a manner spread on the (m+k)-th frame. Interpolation images can be formed without clearance and overlapping by forming each interpolation image in correspondence with each block.
  • The interpolation image can be formed on a frame between the m-th frame 1 and the (m+i)-th frame 3 as described before, and also on a frame between the (m+i)-th frame 3 and the (m+n)-th frame 2 using the same motion vector D. Assuming that i=1 and n=2, for example, a (m+0.5)-th frame image (between the m-th frame 1 and the (m+1)-th frame) and a (m+1.5)-th frame image (between the (m+1)-th frame and the (m+2)-th frame) are approximately simultaneously obtained using the motion vector D detected from the m-th frame 1 and the (m+2)-th frame. This can shorten computation time in comparison with a system which computes a motion vector between m-th frame 1 and the (m+1)-th frame, forms a (m+0.5)-th frame using the motion vector, computes another motion vector between the (m+1)-th frame and the (m+2)-th frame, and forms a (m+1.5)-th frame using the another motion vector.
  • Further, when computation process affords in processing time, the following method is available. At first the (m+0.5)-th frame image and (m+1.5)-th frame image are computed using the motion vector D1 detected by the m-th frame 1 and (m+2)-th frame, and then the (m+1.5)-th frame image and (m+2.5)-th frame image are computed using another motion vector D2 detected by the (m+1)-th frame and (m+3)-th frame. Furthermore, a correlation between the (m+1.5)-th frame image and the (m+2)-th frame image is inspected, and each block is selected so that a correlation with respect to the (m+2)-th frame image becomes large.
  • The Seventh Embodiment
  • A method of detecting a motion vector and forming an interpolation image when an original picture is an interlaced picture will be described referring to FIGS. 21-23.
  • At first the first block 11 is extracted from the motion vector search area of the m-th field 1 (step S701). A plurality of second blocks 12 having a large correlation with respect to the first block are extracted from a plurality of blocks of the same block size as the first block 11 in the motion vector search area of the (m+2n)-th field (n is an integer more than 1) (step S702). The first vector connecting between the first block 11 and the second block 12 is detected as the motion vector D (step S703).
  • The third block (interlaced block) 13 at a spatial position shifted spatially by ½ line in a vertical direction (upper or lower direction) from the first block 11 of the m-th frame 1 is extracted from a (m+2n−1)-th field (step S704). A second motion vector D/2 whose direction is the same as that of the first motion vector D and whose quantity of movement is ½ that of the first motion vector D (step S705) is computed. The fourth block (interlaced block) 14 at a movement position of a block that is located in spatially the same position as that of the first block 11 and moves to the (m+2n)-th field is extracted from the (m+2n)-th field according to the second motion vector D/2 (step S706).
  • A thirteenth block 23 is formed by subjecting the third block 13 to linear interpolation in the first vertical direction (e.g., upper direction)(step S707). Furthermore, a 14th block 24 is formed by subjecting the fourth block 14 to linear interpolation in the second vertical direction (lower direction) opposite to the first vertical direction (step S708). Propriety of the first motion vector 11 is determined using the 13th block 23 and the 14th block 24. In other words, a correlation between, for example, the 13th block 23 and the 14th block 24 is inspected by computing an absolute value difference sum, and the first motion vector D that maximizes the correlation is selected as an optimum motion vector (step S709).
  • An interpolation image of the (m+k)-th field is formed using basically the same method as the sixth embodiment. In other words, similar to steps S602-S604 in FIG. 20 in the sixth embodiment, the third motion vector is obtained by scale-converting the optimum motion vector pursued with step S709 according to a time position of the (m+k)-th field (step S710). The 15th block corresponding to a movement position of the first block 11 from the (m+2n)-th field is extracted from the (m+2n)-th field according to the third motion vector after scale conversion (step S711). The interpolation image of the (m+k)-th field is formed by assigning the 15th block to spatially the same position as the first block 11 of the (m+k)-th field (step S712). The 15th block is an image on the frame (m+n)-th frame. However, when no image exists at the position indicated by the motion vector, that is, an image exists between lines, an image is formed using upper and lower lines. The 15th block is formed by subjecting the upper and lower lines to, for example, a linear interpolation.
  • In k=0.5, for example, when an interpolation image of (m+1.5)-th field is formed, the fifth block corresponding to a movement position to the (m+2n)-th field is extracted from the (m+2n)-th field according to the third motion vector whose direction is the same as the optimum motion vector and whose quantity of movement is ¼ of the first motion vector. The fifth block is assigned to the (m+1.5)-th Field. Similarly, the interpolation image is completed by acquiring the fifth blocks on (m+1.5)-th field and assigning them to the (m+1.5)-th field. This method acquires a progressive image as a field image basically. When the field of interpolation image is subjected to an interlacing process, the field must be vertically skipped with respect to lines.
  • The Eighth Embodiment
  • There will now be described an image display system using the motion vector detection method and the interpolation image forming method of the above embodiments as the eighth embodiment of the present invention.
  • FIG. 24 shows a schematic configuration of an image display system. An input image signal 101 is input to an interpolation field image forming unit 102 and an image switching unit 104. The interpolation field image forming unit 102 forms an interpolation image signal according to the procedure described above. The interpolation image signal 103 is output to the image switching unit 104. The image switching unit 104 switches from the input image signal 101 to the interpolation video signal 104 or vice versa. The output video signal 105 from the image switching unit 104 is output to a high-speed refresh display 106 that is a hold type display. The display 106 changes a refresh rate according to a synchronizing signal included in the output video signal 105, and displays an image.
  • As described above, according to the present invention, the error motion vector different from actual movement of the image is removed and the precision motion vector can be detected. Further, an output image signal of high frame frequency can be formed from an input image signal of low frame frequency by forming an interpolation image using the precision motion vector. A moving image that is more real and a still picture of a little picture quality degradation can be provided.
  • Additional advantages and modifications will readily occur to those skilled in the art. Therefore, the invention in its broader aspects is not limited to the specific details and representative embodiments shown and described herein. Accordingly, various modifications may be made without departing from the spirit or scope of the general inventive concept as defined by the appended claims and their equivalents.

Claims (4)

1. A method of forming an interpolation image at a time position of a (m+k)-th field (m is an integer and k is an real number) between a m-th field of an original image and a (m+2)-th field, the method comprising:
extracting a first block from a motion vector search area of a m-th field (m is an integer);
extracting, from a motion vector search area of a (m+2)-th field, a plurality of second blocks selected in order of decreasing correlation with respect to the first block;
detecting a plurality of first motion vectors between the first block and the plurality of second blocks;
extracting, from a (m+1)-th field, a third block located at a spatial position obtained by shifting spatially the first block in a vertical direction by ½ line;
calculating a plurality of second motion vectors corresponding to ½ times the first motion vectors;
extracting, from the (m+2)-th field, a plurality of fourth blocks located at a destination of the third block located at same spatial position as that of the first block in the (m+1)-th field, according to the second motion vectors;
interpolating linearly the third block in a first vertical direction to create a fifth block;
interpolating linearly the fourth blocks in a second vertical direction opposite to the first vertical direction to create a plurality of sixth blocks;
selecting, from the plurality of first motion vectors, as an optimum motion vector a motion vector corresponding to a maximum one of correlations between the fifth block and the sixth blocks;
scaling the optimum motion vector according to a time position of the (m+k)-th field;
extracting, from the (m+2)-th field, a seventh block located at a destination of an interpolation block located at same spatial position as that of the first block in the (m+k)-th field, according to the scaled optimum motion vector; and
allocating the seventh block to the interpolation block located at the same spatial position as that of the first block in the (m+k)-th field.
2. A computer system comprising:
an interpolation image forming unit configured to form the interpolation image from an original image by using the method of forming the interpolation image according to claim 1; and
a displaying unit configured to display the original image and the interpolation image.
3. A display apparatus comprising:
an interpolation image forming unit configured to form the interpolation image from an original image by using the method of forming the interpolation image according to claim 1; and
a displaying unit configured to display the original image and the interpolation image.
4. A method of displaying an original image and the interpolation image formed by the method of forming the interpolation image according to claim 1.
US11/686,479 2002-03-15 2007-03-15 Motion vector detection method and apparatus Abandoned US20070153904A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/686,479 US20070153904A1 (en) 2002-03-15 2007-03-15 Motion vector detection method and apparatus

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2002-073207 2002-03-15
JP2002073207A JP3840129B2 (en) 2002-03-15 2002-03-15 Motion vector detection method and apparatus, interpolation image generation method and apparatus, and image display system
US10/383,631 US7349475B2 (en) 2002-03-15 2003-03-10 Motion vector detection method and apparatus
US11/686,479 US20070153904A1 (en) 2002-03-15 2007-03-15 Motion vector detection method and apparatus

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US10/383,631 Division US7349475B2 (en) 2002-03-15 2003-03-10 Motion vector detection method and apparatus

Publications (1)

Publication Number Publication Date
US20070153904A1 true US20070153904A1 (en) 2007-07-05

Family

ID=28035227

Family Applications (4)

Application Number Title Priority Date Filing Date
US10/383,631 Expired - Fee Related US7349475B2 (en) 2002-03-15 2003-03-10 Motion vector detection method and apparatus
US11/686,479 Abandoned US20070153904A1 (en) 2002-03-15 2007-03-15 Motion vector detection method and apparatus
US11/686,424 Abandoned US20070165719A1 (en) 2002-03-15 2007-03-15 Motion vector detection method and apparatus
US11/686,457 Expired - Fee Related US8144777B2 (en) 2002-03-15 2007-03-15 Motion vector detection method and apparatus

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US10/383,631 Expired - Fee Related US7349475B2 (en) 2002-03-15 2003-03-10 Motion vector detection method and apparatus

Family Applications After (2)

Application Number Title Priority Date Filing Date
US11/686,424 Abandoned US20070165719A1 (en) 2002-03-15 2007-03-15 Motion vector detection method and apparatus
US11/686,457 Expired - Fee Related US8144777B2 (en) 2002-03-15 2007-03-15 Motion vector detection method and apparatus

Country Status (2)

Country Link
US (4) US7349475B2 (en)
JP (1) JP3840129B2 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080069221A1 (en) * 2006-09-19 2008-03-20 Kabushiki Kaisha Toshiba Apparatus, method, and computer program product for detecting motion vector and for creating interpolation frame
US20090225227A1 (en) * 2008-03-05 2009-09-10 Panasonic Corporation Motion vector detecting device
US20100315550A1 (en) * 2009-06-12 2010-12-16 Masayuki Yokoyama Image frame interpolation device, image frame interpolation method, and image frame interpolation program
US20120215524A1 (en) * 2009-10-26 2012-08-23 Panasonic Corporation Tone determination device and method
US8319889B2 (en) 2009-10-08 2012-11-27 JVC Kenwood Corporation Frame rate conversion apparatus and method
US8433143B1 (en) * 2012-01-04 2013-04-30 Doug Carson & Associates, Inc. Automated detection of video artifacts in an information signal

Families Citing this family (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20050049680A (en) 2003-11-22 2005-05-27 삼성전자주식회사 Noise reduction and de-interlacing apparatus
JP4271115B2 (en) * 2004-09-29 2009-06-03 株式会社東芝 Image matching apparatus, image matching method, and image matching program
JP2006260527A (en) * 2005-02-16 2006-09-28 Toshiba Corp Image matching method and image interpolation method using same
JP4398925B2 (en) * 2005-03-31 2010-01-13 株式会社東芝 Interpolation frame generation method, interpolation frame generation apparatus, and interpolation frame generation program
JP4064413B2 (en) * 2005-06-27 2008-03-19 株式会社東芝 Communication support device, communication support method, and communication support program
JP4398919B2 (en) * 2005-08-22 2010-01-13 株式会社東芝 Image matching apparatus, image matching method, and image matching program
JP4417918B2 (en) * 2006-03-30 2010-02-17 株式会社東芝 Interpolation frame creation device, motion vector detection device, interpolation frame creation method, motion vector detection method, interpolation frame creation program, and motion vector detection program
CN101087405A (en) * 2006-06-07 2007-12-12 鸿富锦精密工业(深圳)有限公司 Dynamic monitoring system and method
KR100767503B1 (en) 2006-06-13 2007-10-17 주식회사 팬택앤큐리텔 Motion vector compensation system and method using the 3-dimension fitting technique
JP4375452B2 (en) * 2007-07-18 2009-12-02 ソニー株式会社 Image processing apparatus, image processing method, program, and display apparatus
JP2009071809A (en) 2007-08-20 2009-04-02 Panasonic Corp Video display apparatus, and interpolated image generating circuit and its method
US20090207314A1 (en) * 2008-02-14 2009-08-20 Brian Heng Method and system for motion vector estimation using a pivotal pixel search
JP4883029B2 (en) * 2008-03-05 2012-02-22 パナソニック株式会社 Motion vector detection circuit, motion vector detection device, and integrated circuit
JP5200788B2 (en) * 2008-09-09 2013-06-05 富士通株式会社 Video signal processing apparatus, video signal processing method, and video signal processing program
US8848110B2 (en) * 2010-03-29 2014-09-30 Silicon Laboratories Inc. Mixed-mode receiver circuit including digital gain control
JP2012089986A (en) 2010-10-18 2012-05-10 Mitsubishi Electric Corp Image processing device and method, and image display device and method
JP4735779B2 (en) * 2011-01-12 2011-07-27 日本ビクター株式会社 Interpolated pixel data generation apparatus and method
JP5867289B2 (en) * 2012-05-23 2016-02-24 株式会社Jvcケンウッド Motion vector generation apparatus and method
US10205962B2 (en) * 2013-03-15 2019-02-12 Raymond Zenkich System and method for non-uniform video coding
EP3820138A1 (en) * 2019-11-06 2021-05-12 Koninklijke Philips N.V. A system for performing image motion compensation

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4496972A (en) * 1980-05-10 1985-01-29 Deutsche Forschungs-Und Versuchsanstalt Fur Luft-Und Raumfahrt E.V. Method for the representation of video images or scenes, in particular aerial images transmitted at reduced frame rate
US4651207A (en) * 1984-03-05 1987-03-17 Ant Nachrichtentechnik Gmbh Motion adaptive interpolation of television image sequences
US5418570A (en) * 1992-03-03 1995-05-23 Kabushiki Kaisha Toshiba Motion picture coding apparatus
US5610665A (en) * 1993-10-12 1997-03-11 Berman; John L. Interactive television graphics interface
US5822007A (en) * 1993-09-08 1998-10-13 Thomson Multimedia S.A. Method and apparatus for motion estimation using block matching
US6005639A (en) * 1990-09-20 1999-12-21 Bbc, Broadcasting House Vector assignment for video image motion compensation
US6008865A (en) * 1997-02-14 1999-12-28 Eastman Kodak Company Segmentation-based method for motion-compensated frame interpolation
US6351545B1 (en) * 1999-12-14 2002-02-26 Dynapel Systems, Inc. Motion picture enhancing system
US20020036705A1 (en) * 2000-06-13 2002-03-28 Samsung Electronics Co., Ltd. Format converter using bi-directional motion vector and method thereof
US6594313B1 (en) * 1998-12-23 2003-07-15 Intel Corporation Increased video playback framerate in low bit-rate video applications
US6618439B1 (en) * 1999-07-06 2003-09-09 Industrial Technology Research Institute Fast motion-compensated video frame interpolator

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE3663875D1 (en) 1986-03-08 1989-07-13 Ant Nachrichtentech Motion compensating field interpolation method using a hierarchically structured displacement estimator
GB2231752B (en) 1989-04-27 1993-08-04 Sony Corp Motion dependent video signal processing
JPH0683441B2 (en) 1989-06-26 1994-10-19 松下電器産業株式会社 Method and apparatus for inter-frame interpolation coding of image signal
JPH089340A (en) 1994-06-24 1996-01-12 Mitsubishi Electric Corp Motion vector detector and motion vector detection method
JPH09214899A (en) 1996-01-31 1997-08-15 Nippon Hoso Kyokai <Nhk> Image prediction processing method an its device
JP3774028B2 (en) 1997-04-23 2006-05-10 日本放送協会 Motion estimation apparatus for moving images
US6366317B1 (en) * 1998-03-27 2002-04-02 Intel Corporation Motion estimation using intrapixel logic

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4496972A (en) * 1980-05-10 1985-01-29 Deutsche Forschungs-Und Versuchsanstalt Fur Luft-Und Raumfahrt E.V. Method for the representation of video images or scenes, in particular aerial images transmitted at reduced frame rate
US4651207A (en) * 1984-03-05 1987-03-17 Ant Nachrichtentechnik Gmbh Motion adaptive interpolation of television image sequences
US6005639A (en) * 1990-09-20 1999-12-21 Bbc, Broadcasting House Vector assignment for video image motion compensation
US5418570A (en) * 1992-03-03 1995-05-23 Kabushiki Kaisha Toshiba Motion picture coding apparatus
US5822007A (en) * 1993-09-08 1998-10-13 Thomson Multimedia S.A. Method and apparatus for motion estimation using block matching
US5610665A (en) * 1993-10-12 1997-03-11 Berman; John L. Interactive television graphics interface
US6008865A (en) * 1997-02-14 1999-12-28 Eastman Kodak Company Segmentation-based method for motion-compensated frame interpolation
US6594313B1 (en) * 1998-12-23 2003-07-15 Intel Corporation Increased video playback framerate in low bit-rate video applications
US6618439B1 (en) * 1999-07-06 2003-09-09 Industrial Technology Research Institute Fast motion-compensated video frame interpolator
US6351545B1 (en) * 1999-12-14 2002-02-26 Dynapel Systems, Inc. Motion picture enhancing system
US20020036705A1 (en) * 2000-06-13 2002-03-28 Samsung Electronics Co., Ltd. Format converter using bi-directional motion vector and method thereof

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080069221A1 (en) * 2006-09-19 2008-03-20 Kabushiki Kaisha Toshiba Apparatus, method, and computer program product for detecting motion vector and for creating interpolation frame
US8189104B2 (en) * 2006-09-19 2012-05-29 Kabushiki Kaisha Toshiba Apparatus, method, and computer program product for detecting motion vector and for creating interpolation frame
US20090225227A1 (en) * 2008-03-05 2009-09-10 Panasonic Corporation Motion vector detecting device
US20100315550A1 (en) * 2009-06-12 2010-12-16 Masayuki Yokoyama Image frame interpolation device, image frame interpolation method, and image frame interpolation program
US8319889B2 (en) 2009-10-08 2012-11-27 JVC Kenwood Corporation Frame rate conversion apparatus and method
US20120215524A1 (en) * 2009-10-26 2012-08-23 Panasonic Corporation Tone determination device and method
US8670980B2 (en) * 2009-10-26 2014-03-11 Panasonic Corporation Tone determination device and method
US8433143B1 (en) * 2012-01-04 2013-04-30 Doug Carson & Associates, Inc. Automated detection of video artifacts in an information signal

Also Published As

Publication number Publication date
US20070165719A1 (en) 2007-07-19
JP2003274414A (en) 2003-09-26
US20030174777A1 (en) 2003-09-18
US20070153903A1 (en) 2007-07-05
US7349475B2 (en) 2008-03-25
US8144777B2 (en) 2012-03-27
JP3840129B2 (en) 2006-11-01

Similar Documents

Publication Publication Date Title
US8144777B2 (en) Motion vector detection method and apparatus
JP4157579B2 (en) Image display apparatus and method, image processing apparatus and method
US7180548B2 (en) Method of generating frame interpolation image and an apparatus therefor
US7561621B2 (en) Method of searching for motion vector, method of generating frame interpolation image and display system
US7412114B2 (en) Method of generating an interpolation image, an interpolation image generating apparatus, and an image display system using the same
US8189104B2 (en) Apparatus, method, and computer program product for detecting motion vector and for creating interpolation frame
US20070140529A1 (en) Method and device for calculating motion vector between two images and program of calculating motion vector between two images
US6925124B2 (en) Unit for and method of motion estimation and image processing apparatus provided with such motion estimation unit
JP4355347B2 (en) Image display apparatus and method, image processing apparatus and method
JP4082664B2 (en) Video search device
JP2007060192A (en) Interpolating frame generator, its method, image display system and program and recording medium
US20050002456A1 (en) Motion vector detector for frame rate conversion and method thereof
JP2006215657A (en) Method, apparatus, program and program storage medium for detecting motion vector
US7143351B1 (en) Image display device and displaying method
US6707502B1 (en) Apparatus and method for converting a field frequency of a picture signal
US20030012283A1 (en) Motion vector detecting device and self-testing method therein
JP4181189B2 (en) Motion vector detection method and apparatus, interpolation image generation method and apparatus, and image display system
JP3189655B2 (en) Apparatus and method for detecting motion vector
JP4157587B2 (en) Image display apparatus and method, image processing apparatus and method
JPH08242454A (en) Method for detecting global motion parameter
US20030086497A1 (en) Motion vector detecting device improved in detection speed of motion vectors and system employing the same devices
JP2008109628A (en) Image display apparatus and method, image processor and method
JP4181190B2 (en) Motion vector detection method and apparatus, interpolation image generation method and apparatus, and image display system
JPH11187424A (en) Detector and detection method for motion vector detector and motion vector detection method
JP4157586B2 (en) Image display apparatus and method, image processing apparatus and method

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION