US5737449A - Apparatus for encoding a contour of regions contained in a video signal - Google Patents

Apparatus for encoding a contour of regions contained in a video signal Download PDF

Info

Publication number
US5737449A
US5737449A US08/580,237 US58023795A US5737449A US 5737449 A US5737449 A US 5737449A US 58023795 A US58023795 A US 58023795A US 5737449 A US5737449 A US 5737449A
Authority
US
United States
Prior art keywords
points
contour
current
vertex
previous
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US08/580,237
Inventor
Min-Sup Lee
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Quarterhill Inc
Original Assignee
Daewoo Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Daewoo Electronics Co Ltd filed Critical Daewoo Electronics Co Ltd
Assigned to DAEWOO ELECTRONICS, CO., LTD reassignment DAEWOO ELECTRONICS, CO., LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LEE, MIN-SUP
Application granted granted Critical
Publication of US5737449A publication Critical patent/US5737449A/en
Assigned to DAEWOO ELECTRONICS CORPORATION reassignment DAEWOO ELECTRONICS CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: DAEWOO ELECTRONICS CO., LTD.
Assigned to MAPLE VISION TECHNOLOGIES INC. reassignment MAPLE VISION TECHNOLOGIES INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: DAEWOO ELECTRONICS CORPORATION
Anticipated expiration legal-status Critical
Assigned to QUARTERHILL INC. reassignment QUARTERHILL INC. MERGER AND CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: MAPLE VISION TECHNOLOGIES INC., QUARTERHILL INC.
Assigned to WI-LAN INC. reassignment WI-LAN INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: QUARTERHILL INC.
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/20Contour coding, e.g. using detection of edges

Definitions

  • the present invention relates to an apparatus for encoding a video signal; and, more particularly, to an apparatus capable of effectively encoding a contour of regions contained in a video signal, thereby reducing the amount of data to be transmitted.
  • One of such methods for encoding video signals for a low bit-rate encoding system is the so-called region-based coding technique employing a region-based segmentation approach.
  • a contour for the segmented regions is extracted or defined by using a feature extraction process, wherein the contour provides boundary or contour information for tracing the boundary of the segmented regions, the boundary information including position data of the pixels along the boundary of the regions in the image.
  • the extracted contour may be then coded by using various coding techniques.
  • the texture information representing a mean gray level of the pixels included in each of the segmented regions is coded.
  • a classical coding method for representing the contour information is chain coding.
  • the chain coding method requires a substantial amount of bits for the representation thereof, although there is no loss in the contour information.
  • contour information such as polygonal approximation and B-spline approximation.
  • One of the disadvantages in the polygonal approximation is the rough representation of the contour.
  • the B-spline approximation is, on the other hand, capable of representing the contour more precisely; however, it requires a high-order polynomial to reduce the approximation error, thereby increasing the overall computational complexity of the video encoder.
  • DST discrete sine transform
  • a number of vertex points are determined and the contour of an object is approximated through the use of a polygonal approximation for fitting the contour by line segments.
  • N sample points for each line segment are selected and an approximation error at each of the N sample points located on each line segment is sequentially calculated in order to obtain a set of approximation errors for each line segment.
  • the N sample points are equi-spaced on each line segment and each of the approximation errors represents the distance between each of the N sample points and the contour.
  • sets of DST coefficients are generated by performing one-dimensional DST operation on each set of approximation errors.
  • an apparatus for encoding a contour of regions contained in a video signal having a plurality of frames including a current frame and a previous frame, which comprises: a first segmentation block for segmenting the previous frame into a number of regions to generate a previous contour; a second segmentation block for segmenting the current frame into a number of regions to generate a current contour; a prediction parameter calculation block for generating prediction parameters based on the previous and the current contours, and the previous and the current frames; a contour prediction block for constructing a predicted current contour by transforming the previous contour based on the prediction parameters; an error detection block for detecting deviation points lying on portions of the predicted current contour that deviate from the current contour, to thereby generate the deviation points and corresponding displacements to the deviation points, wherein each of the displacements represents a minimum distance in a horizontal direction between each of the deviation points and the current contour; and a contour coder for encoding the deviation points and the displacements together with the prediction parameters.
  • FIG. 1 represents a block diagram of a video signal encoding apparatus in accordance with a preferred embodiment of the present invention
  • FIG. 2 presents a detailed block diagram of the prediction parameter calculation block shown in FIG. 1;
  • FIG. 3 shows a detailed block diagram of the error detection block shown in FIG. 1;
  • FIG. 4 describes a technique for selecting vertex points through the use of grids and contours
  • FIG. 5 illustrates an exemplary diagram for explaining a contour reconstruction process
  • FIGS. 6A and 6B provide exemplary diagrams, each representing errors between a line segment joining two relocated deviation points and its corresponding current contour.
  • FIG. 1 there is shown a block diagram of the inventive apparatus for encoding a contour of regions contained in a video signal.
  • An input video signal is fed, as a current frame signal, to a second segmentation block 16 and a prediction parameter calculation block 18 through a line L1; and is stored as a previous frame signal in a frame memory 12 which is connected to a first segmentation block 14 and the prediction parameter calculation block 18 through a line L2.
  • the first segmentation block 14 segments the previous frame retrieved from the frame memory 12 into a number of regions and generates a previous contour onto a line wherein the previous contour provides previous boundary information for tracing the boundary of the segmented regions of the previous frame, the previous boundary information including position data of the pixels along the boundary of the regions in the previous frame.
  • the contour data representing the contour for the segmented regions on the line L3 from the first segmentation block 14 is provided to the prediction parameter calculation block 18 and a contour prediction block 20.
  • the second segmentation block 16 performs the same function as the first segmentation block 14, to thereby generate a current contour to the prediction parameter calculation block 18 and an error detection block 22. That is, the second segmentation block 16 segments the input video signal supplied thereto as the current frame signal into a number of regions and generates a current contour onto a line L4, wherein the current contour provides current boundary information for tracing the boundary of the segmented regions of the current frame, the current boundary information including position data of the pixels along the boundary of the regions in the current frame.
  • the contour data representing the contour for the segmented regions on the line L4 from the second segmentation block 16 is provided to the prediction parameter calculation block 18 and the error detection block 22.
  • the prediction parameter calculation block 18 of the present invention determines a predetermined number of vertex points on the previous contour to produce vertex information representing the positions of the vertex points of the previous contour; and detects predicted vertex points based on a similarity calculation between the current and the previous frames through the use of vertex information and the current boundary information from the current contour to provide predicted vertex information representing the positions of the predicted vertex points. Thereafter, a known affine transform algorithm is carried out based on the vertex point information and the predicted vertex information, to thereby generate affine transform coefficients as prediction parameters.
  • the prediction parameters are coupled through a line L5 to the contour prediction block 20 and a contour coder 24. Details of the prediction parameter calculation block 18 will be described hereinafter with reference to FIGS. 2 and 4.
  • the contour prediction block 20 constructs a predicted current contour by affine transformation of the previous contour based on the prediction parameters.
  • the predicted current contour from the contour prediction block 20 is coupled through a line L6 to the error detection block 22.
  • the error detection block 22 selects a number of points from the points on the predicted current contour as feature points by using a grid, and determines deviation points among the feature points, lying on portions of the predicted current contour that deviate from the current contour; and calculates a displacement for each of the deviation points, to thereby generate the deviation points, the displacements for the deviation points and a grid index, wherein each of the displacements represents a minimum distance in a horizontal direction between each of the deviation points and the current contour.
  • the deviation points, the displacements for the deviation points and the grid index are then transmitted through a line L7 to the contour coder 24 for further processing. More details of the error detection block 22 will be described hereinafter with reference to FIGS. 3 and 5.
  • the deviation points and the displacements for the deviation points are encoded together with the prediction parameters and the grid index by using, e.g., a binary arithmetic code of JPEG(Joint Photographic Experts Group).
  • the encoded digital signal comprising the encoded deviation points, the displacements, the prediction parameters and the grid index is sent to a transmitter(not shown) for the transmission thereof.
  • the prediction parameter calculation block 18 includes a vertex point selection block 101, a block formation section 103, a candidate block generation section 105, a vertex prediction block 107 and a transform coefficient determinater 109.
  • the previous contour on the line L3 from the first segmentation block 14 is inputted to the vertex point selection block 101.
  • the vertex point selection block 101 determines a predetermined number of points from the points on the previous contour as vertex points to generate vertex information representing the positions of the vertex points of the previous contour.
  • the vertex points are determined by using a known grid technique, which employs one of the various grid types, e.g., a rectangular grid, together with an edge detection technique as shown in FIG. 4. In this scheme, intersection points of the grid and the contour are selected as vertex points.
  • the vertex information of the selected vertex points from the vertex point selection block 101 are inputted to the block formation section 103 and the transform coefficient determinater 109.
  • a vertex point block of, e.g., 8 ⁇ 8 pixels of the previous frame, having each of the vertex points at the center thereof, is retrieved via the line L2 from the frame memory 12 and provided to the vertex prediction block 107.
  • the input video signal on the line L1 is fed as the current frame signal to the candidate block generation section 105 and the current contour on the line L4 is also fed to the candidate block generation section 105.
  • the candidate block generation section 105 defines a generally larger search region than the vertex point block of the current frame and generates therefrom a plurality of equalsized candidate blocks to the vertex prediction block 107.
  • the candidate blocks defined in the search region are chosen such that their center pixels form a part of the current contour, thereby placing the predicted vertex points on the current contour.
  • a motion vector for each of the vertex point blocks from the block formation section 103 is determined after a similarity calculation between the vertex point block and each of the plurality of candidate blocks included in the search region from the candidate block generation section 105, wherein each of the motion vectors represents a spatial displacement between a vertex point block in the previous frame and a most similar candidate block thereto in the current frame.
  • the determined motion vector for the vertex point block is assigned as the motion vector of the vertex point contained in that vertex point blocks
  • predicted vertex points are estimated based on the motion vectors for the vertex points and the predicted vertex information representing the positions of the predicted vertex points of the current contour are generated to the transform coefficient determinater 109.
  • affine transform coefficients are determined by using an optimum motion parameter determination technique, disclosed in a commonly owned copending application, U.S. Ser. No. 08/502,228, entitled “METHOD FOR SEGMENTING AND ESTIMATING A MOVING OBJECT MOTION USING A HIERARCHY OF MOTION MODELS", based on the vertex information from the vertex point selection block 101 and the predicted vertex information from the vertex prediction block 107.
  • the determined affine transform coefficients are then provided through the line L5 to the contour prediction block 20 and the contour coder 24 as the prediction parameters.
  • FIG. 3 depicts a detailed block diagram of the error detection block 22 shown in FIG. 1.
  • the error detection block 22 is provided with a feature point selection block 202, a displacement detection block 204, a contour reconstruction block 206, a similarity measurement block 208, a grid generation block 210 and a switching block 212.
  • the grid generation block 210 serves to provide the feature point selection block 202 with grid information corresponding to one of a plurality of grids selected therein in response to each grid generation signal sequentially provided from the similarity measurement block 208, wherein each of the grids includes a number of grid lines, grid line densities for the grids being different from each other.
  • a grid selected at the grid generation block 210 in response to a grid generation signal has a lower grid line density than a grid corresponding to a next grid generation signal.
  • the current contour on the line L4 from the second segmentation block 16 is fed to the similarity measurement block 208.
  • the similarity measurement block 208 Upon receiving the current contour, the similarity measurement block 208 generates a first grid generation signal to the grid generation block 210.
  • the grid generation block 210 in response to the first grid generation signal, selects, among the plurality of grids, a grid having a lowest grid line density to provide the feature point selection block 202 with grid information therefor.
  • the predicted current contour on the line L6 from the contour prediction block 20 is inputted to the feature point selection block 202, the displacement detection block 204 and the contour reconstruction block 206.
  • the feature point selection block 202 determines a number of feature points on the predicted current contour by using the grid represented by the grid information provided from the grid generation block 210, wherein the feature points represent intersection points between the grid and the predicted current contour.
  • the feature points selected from the feature point selection block 202 is fed to the displacement detection block 204.
  • the current contour on the line L4 from the second segmentation block 16 is also inputted to the displacement detection block 204.
  • deviation points are determined among the feature points, lying on portions of the predicted current contour that deviate from the current contour, and the displacement for each of the deviation points is calculated.
  • the deviation points and the displacements therefor are provided from the displacement detection block 204 to the contour reconstruction block 206 and the switching block 212.
  • the contour reconstruction block 206 relocates the deviation points on the predicted current contour onto the current contour based on the displacement for each of the deviation points and generates a reconstructed current contour by joining the relocated deviation points on the current contour.
  • the reconstructed current contour is provided from the contour reconstruction block 206 to the similarity measurement block 208 which also receives the current contour on the line L4 from the second segmentation block 16.
  • FIG. 5 illustrates an exemplary diagram for explaining a contour reconstruction process.
  • deviation points A to D on the predicted current contour C' are relocated onto the current contour C based on the displacement for each of the deviation points. Thereafter, the relocated deviation points A' to D' on the current contour C are joined together, thereby producing the reconstructed current contour.
  • the similarity measurement block 208 provides N sample points for each line segment joining two neighboring relocated deviation points; and calculates an error for each of the N sample points on each of the line segments to produce a set of errors for each of the line segments, wherein the N sample points are equi-spaced on each of the line segments and each of the errors represents a distance between each of the N sample points and the current contour.
  • FIGS. 6A and 6B illustrate exemplary diagrams representing errors between line segments and corresponding contour segments. Specifically, FIG. 6A depicts errors between the line segment A'B' and its corresponding contour segment and FIG. 6B shows errors between the line segment B'C' and its corresponding contour segment.
  • Each of the errors d1 to d4 or d1' to d4' represents the distance from each of the sample points S1 to S4 on the line segment A'B' or S1' to S4' on the line segment B'C' to the corresponding contour segment.
  • the errors for the relocated deviation points are all "zeros". For all the relocated deviation points are positioned on the current contour.
  • the similarity calculation block 208 sums up the set of errors in order to provide a similarity index. If the similarity index is larger than or equal to a predetermined threshold value, the similarity measurement block 208 generates a next, e.g., second grid generation signal to the grid generation block 210; otherwise, it generates a completion signal to the grid generation block 210 and the switching block 212. The switching block 212 relays the deviation points and the displacements to the contour coder 24 in response to the completion signal from the similarity measurement block 208.
  • the grid generation block 210 selects, in response to the next grid generation signal, a grid of a next lower grid line density and provides grid information therefor to the feature point selection block 202; and the error detection process described above is repeated with respect to the newly selected grid. On the other hand, if the completion signal is received, the grid generation block 210 generates to the contour coder 24 a grid index indicating the grid which yields the similarity index smaller than the predetermined threshold value.
  • the present invention is capable of substantially reducing the volume of transmission data for representing the contour of regions contained in a video signal by transmitting the prediction parameters together with the deviation information including the deviation points and the displacements for the deviation points.

Abstract

An apparatus for encoding a contour of regions contained in a video signal comprises first and second segmentation blocks for segmenting previous and current frame into a number of regions to generate previous and current contours, a prediction parameter calculation block for generating prediction parameters based on the previous and current contours, and the previous and current frames, a contour prediction block for constructing a predicted current contour by transforming the previous contour based on the prediction parameters, an error detection block for detecting deviation points lying on portions of the predicted current contour that deviate from the current contour, thereby generating the deviation points and corresponding displacements to the deviation points, and a contour coder for encoding the deviation points and the displacements together with the prediction parameters.

Description

FIELD OF THE INVENTION
The present invention relates to an apparatus for encoding a video signal; and, more particularly, to an apparatus capable of effectively encoding a contour of regions contained in a video signal, thereby reducing the amount of data to be transmitted.
DESCRIPTION OF THE PRIOR ART
In digital video systems such as video-telephone, teleconference and high definition television, a large amount of digital data is needed to define each video frame signal since a video line signal in the video frame signal comprises a sequence of digital data referred to as pixel values. Since, however, the available frequency bandwidth of a conventional transmission channel is limited, in order to transmit the substantial amount of digital data therethrough, it is inevitable to compress or reduce the volume of data through the use of various data compression techniques, especially, in the case of such low bit-rate video signal codec(coding-decoding) systems as video-telephone and teleconference apparatus.
One of such methods for encoding video signals for a low bit-rate encoding system is the so-called region-based coding technique employing a region-based segmentation approach.
In the region-based coding technique, various regions in an image that have similar features are identified, and similar adjacent regions are merged sequentially until the remaining adjacent regions become sufficiently different, thereby producing a limited number of distinct segmented regions of the image. Thereafter, a contour for the segmented regions is extracted or defined by using a feature extraction process, wherein the contour provides boundary or contour information for tracing the boundary of the segmented regions, the boundary information including position data of the pixels along the boundary of the regions in the image. The extracted contour may be then coded by using various coding techniques. And, the texture information representing a mean gray level of the pixels included in each of the segmented regions is coded.
A classical coding method for representing the contour information is chain coding. The chain coding method, however, requires a substantial amount of bits for the representation thereof, although there is no loss in the contour information.
Accordingly, there have been proposed several improved methods to encode contour information such as polygonal approximation and B-spline approximation. One of the disadvantages in the polygonal approximation is the rough representation of the contour. The B-spline approximation is, on the other hand, capable of representing the contour more precisely; however, it requires a high-order polynomial to reduce the approximation error, thereby increasing the overall computational complexity of the video encoder.
One of the techniques introduced to ameliorate the problems associated with the rough representation of the contour and the computational complexity in the above approximation approaches is a contour approximation method employing a discrete sine transform (DST).
In an apparatus which adopts the contour approximation technique based on the polygonal approximation and the DST, a number of vertex points are determined and the contour of an object is approximated through the use of a polygonal approximation for fitting the contour by line segments. And, N sample points for each line segment are selected and an approximation error at each of the N sample points located on each line segment is sequentially calculated in order to obtain a set of approximation errors for each line segment. The N sample points are equi-spaced on each line segment and each of the approximation errors represents the distance between each of the N sample points and the contour. Thereafter, sets of DST coefficients are generated by performing one-dimensional DST operation on each set of approximation errors.
Even though it is possible to remedy the rough representation and computational complexity, and reduce the volume of transmission data through the use of the aforementioned DST based contour approximation, it still remains desirable to further reduce the volume of transmission data in order to successfully implement a low-bit rate codec system having, e.g., a 64 kb/s transmission channel bandwidth.
SUMMARY OF THE INVENTION
It is, therefore, a primary object of the invention to provide an improved apparatus for encoding a contour of regions in a video signal, which is capable of further reducing the amount of transmission data.
In accordance with the invention, there is provided an apparatus for encoding a contour of regions contained in a video signal, the video signal having a plurality of frames including a current frame and a previous frame, which comprises: a first segmentation block for segmenting the previous frame into a number of regions to generate a previous contour; a second segmentation block for segmenting the current frame into a number of regions to generate a current contour; a prediction parameter calculation block for generating prediction parameters based on the previous and the current contours, and the previous and the current frames; a contour prediction block for constructing a predicted current contour by transforming the previous contour based on the prediction parameters; an error detection block for detecting deviation points lying on portions of the predicted current contour that deviate from the current contour, to thereby generate the deviation points and corresponding displacements to the deviation points, wherein each of the displacements represents a minimum distance in a horizontal direction between each of the deviation points and the current contour; and a contour coder for encoding the deviation points and the displacements together with the prediction parameters.
BRIEF DESCRIPTION OF THE DRAWINGS
The above and other objects and features of the present invention will become apparent from the following description of preferred embodiments given in conjunction with the accompanying drawings, in which:
FIG. 1 represents a block diagram of a video signal encoding apparatus in accordance with a preferred embodiment of the present invention;
FIG. 2 presents a detailed block diagram of the prediction parameter calculation block shown in FIG. 1;
FIG. 3 shows a detailed block diagram of the error detection block shown in FIG. 1;
FIG. 4 describes a technique for selecting vertex points through the use of grids and contours;
FIG. 5 illustrates an exemplary diagram for explaining a contour reconstruction process; and
FIGS. 6A and 6B provide exemplary diagrams, each representing errors between a line segment joining two relocated deviation points and its corresponding current contour.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
Referring to FIG. 1, there is shown a block diagram of the inventive apparatus for encoding a contour of regions contained in a video signal. An input video signal is fed, as a current frame signal, to a second segmentation block 16 and a prediction parameter calculation block 18 through a line L1; and is stored as a previous frame signal in a frame memory 12 which is connected to a first segmentation block 14 and the prediction parameter calculation block 18 through a line L2.
The first segmentation block 14 segments the previous frame retrieved from the frame memory 12 into a number of regions and generates a previous contour onto a line wherein the previous contour provides previous boundary information for tracing the boundary of the segmented regions of the previous frame, the previous boundary information including position data of the pixels along the boundary of the regions in the previous frame. The contour data representing the contour for the segmented regions on the line L3 from the first segmentation block 14 is provided to the prediction parameter calculation block 18 and a contour prediction block 20.
The second segmentation block 16 performs the same function as the first segmentation block 14, to thereby generate a current contour to the prediction parameter calculation block 18 and an error detection block 22. That is, the second segmentation block 16 segments the input video signal supplied thereto as the current frame signal into a number of regions and generates a current contour onto a line L4, wherein the current contour provides current boundary information for tracing the boundary of the segmented regions of the current frame, the current boundary information including position data of the pixels along the boundary of the regions in the current frame. The contour data representing the contour for the segmented regions on the line L4 from the second segmentation block 16 is provided to the prediction parameter calculation block 18 and the error detection block 22.
The prediction parameter calculation block 18 of the present invention determines a predetermined number of vertex points on the previous contour to produce vertex information representing the positions of the vertex points of the previous contour; and detects predicted vertex points based on a similarity calculation between the current and the previous frames through the use of vertex information and the current boundary information from the current contour to provide predicted vertex information representing the positions of the predicted vertex points. Thereafter, a known affine transform algorithm is carried out based on the vertex point information and the predicted vertex information, to thereby generate affine transform coefficients as prediction parameters. The prediction parameters are coupled through a line L5 to the contour prediction block 20 and a contour coder 24. Details of the prediction parameter calculation block 18 will be described hereinafter with reference to FIGS. 2 and 4. The contour prediction block 20 constructs a predicted current contour by affine transformation of the previous contour based on the prediction parameters. The predicted current contour from the contour prediction block 20 is coupled through a line L6 to the error detection block 22.
In accordance with the present invention, the error detection block 22 selects a number of points from the points on the predicted current contour as feature points by using a grid, and determines deviation points among the feature points, lying on portions of the predicted current contour that deviate from the current contour; and calculates a displacement for each of the deviation points, to thereby generate the deviation points, the displacements for the deviation points and a grid index, wherein each of the displacements represents a minimum distance in a horizontal direction between each of the deviation points and the current contour. The deviation points, the displacements for the deviation points and the grid index are then transmitted through a line L7 to the contour coder 24 for further processing. More details of the error detection block 22 will be described hereinafter with reference to FIGS. 3 and 5.
At the contour coder 24, the deviation points and the displacements for the deviation points are encoded together with the prediction parameters and the grid index by using, e.g., a binary arithmetic code of JPEG(Joint Photographic Experts Group). The encoded digital signal comprising the encoded deviation points, the displacements, the prediction parameters and the grid index is sent to a transmitter(not shown) for the transmission thereof.
Referring now to FIG. 2, there is illustrated a detailed block diagram of the prediction parameter calculation block 18, shown in FIG. 1, for explaining the process for generating prediction parameters. The prediction parameter calculation block 18 includes a vertex point selection block 101, a block formation section 103, a candidate block generation section 105, a vertex prediction block 107 and a transform coefficient determinater 109.
As shown, the previous contour on the line L3 from the first segmentation block 14 is inputted to the vertex point selection block 101. The vertex point selection block 101 determines a predetermined number of points from the points on the previous contour as vertex points to generate vertex information representing the positions of the vertex points of the previous contour. In a preferred embodiment of the present invention, the vertex points are determined by using a known grid technique, which employs one of the various grid types, e.g., a rectangular grid, together with an edge detection technique as shown in FIG. 4. In this scheme, intersection points of the grid and the contour are selected as vertex points. The vertex information of the selected vertex points from the vertex point selection block 101 are inputted to the block formation section 103 and the transform coefficient determinater 109. At the block formation section 103, a vertex point block of, e.g., 8×8 pixels of the previous frame, having each of the vertex points at the center thereof, is retrieved via the line L2 from the frame memory 12 and provided to the vertex prediction block 107.
In the meanwhile, the input video signal on the line L1 is fed as the current frame signal to the candidate block generation section 105 and the current contour on the line L4 is also fed to the candidate block generation section 105. The candidate block generation section 105 defines a generally larger search region than the vertex point block of the current frame and generates therefrom a plurality of equalsized candidate blocks to the vertex prediction block 107. In the preferred embodiment of the invention, the candidate blocks defined in the search region are chosen such that their center pixels form a part of the current contour, thereby placing the predicted vertex points on the current contour.
At the vertex prediction block 107, a motion vector for each of the vertex point blocks from the block formation section 103 is determined after a similarity calculation between the vertex point block and each of the plurality of candidate blocks included in the search region from the candidate block generation section 105, wherein each of the motion vectors represents a spatial displacement between a vertex point block in the previous frame and a most similar candidate block thereto in the current frame. The determined motion vector for the vertex point block is assigned as the motion vector of the vertex point contained in that vertex point blocks Thereafter, predicted vertex points are estimated based on the motion vectors for the vertex points and the predicted vertex information representing the positions of the predicted vertex points of the current contour are generated to the transform coefficient determinater 109. At the transform coefficient determinater 109, affine transform coefficients are determined by using an optimum motion parameter determination technique, disclosed in a commonly owned copending application, U.S. Ser. No. 08/502,228, entitled "METHOD FOR SEGMENTING AND ESTIMATING A MOVING OBJECT MOTION USING A HIERARCHY OF MOTION MODELS", based on the vertex information from the vertex point selection block 101 and the predicted vertex information from the vertex prediction block 107. The determined affine transform coefficients are then provided through the line L5 to the contour prediction block 20 and the contour coder 24 as the prediction parameters.
FIG. 3 depicts a detailed block diagram of the error detection block 22 shown in FIG. 1. The error detection block 22 is provided with a feature point selection block 202, a displacement detection block 204, a contour reconstruction block 206, a similarity measurement block 208, a grid generation block 210 and a switching block 212.
The grid generation block 210 serves to provide the feature point selection block 202 with grid information corresponding to one of a plurality of grids selected therein in response to each grid generation signal sequentially provided from the similarity measurement block 208, wherein each of the grids includes a number of grid lines, grid line densities for the grids being different from each other. A grid selected at the grid generation block 210 in response to a grid generation signal has a lower grid line density than a grid corresponding to a next grid generation signal. The current contour on the line L4 from the second segmentation block 16 is fed to the similarity measurement block 208. Upon receiving the current contour, the similarity measurement block 208 generates a first grid generation signal to the grid generation block 210. The grid generation block 210, in response to the first grid generation signal, selects, among the plurality of grids, a grid having a lowest grid line density to provide the feature point selection block 202 with grid information therefor.
Meantime, the predicted current contour on the line L6 from the contour prediction block 20 is inputted to the feature point selection block 202, the displacement detection block 204 and the contour reconstruction block 206. The feature point selection block 202 determines a number of feature points on the predicted current contour by using the grid represented by the grid information provided from the grid generation block 210, wherein the feature points represent intersection points between the grid and the predicted current contour. The feature points selected from the feature point selection block 202 is fed to the displacement detection block 204. The current contour on the line L4 from the second segmentation block 16 is also inputted to the displacement detection block 204. At the displacement detection block 204, deviation points are determined among the feature points, lying on portions of the predicted current contour that deviate from the current contour, and the displacement for each of the deviation points is calculated. The deviation points and the displacements therefor are provided from the displacement detection block 204 to the contour reconstruction block 206 and the switching block 212. The contour reconstruction block 206 relocates the deviation points on the predicted current contour onto the current contour based on the displacement for each of the deviation points and generates a reconstructed current contour by joining the relocated deviation points on the current contour. The reconstructed current contour is provided from the contour reconstruction block 206 to the similarity measurement block 208 which also receives the current contour on the line L4 from the second segmentation block 16. FIG. 5 illustrates an exemplary diagram for explaining a contour reconstruction process. As shown, deviation points A to D on the predicted current contour C' are relocated onto the current contour C based on the displacement for each of the deviation points. Thereafter, the relocated deviation points A' to D' on the current contour C are joined together, thereby producing the reconstructed current contour.
Referring back to FIG. 3, the similarity measurement block 208 provides N sample points for each line segment joining two neighboring relocated deviation points; and calculates an error for each of the N sample points on each of the line segments to produce a set of errors for each of the line segments, wherein the N sample points are equi-spaced on each of the line segments and each of the errors represents a distance between each of the N sample points and the current contour. FIGS. 6A and 6B illustrate exemplary diagrams representing errors between line segments and corresponding contour segments. Specifically, FIG. 6A depicts errors between the line segment A'B' and its corresponding contour segment and FIG. 6B shows errors between the line segment B'C' and its corresponding contour segment. Each of the errors d1 to d4 or d1' to d4' represents the distance from each of the sample points S1 to S4 on the line segment A'B' or S1' to S4' on the line segment B'C' to the corresponding contour segment. As can be seen from FIGS. 6A and 6B, the errors for the relocated deviation points are all "zeros". For all the relocated deviation points are positioned on the current contour.
In accordance with the invention, the similarity calculation block 208 sums up the set of errors in order to provide a similarity index. If the similarity index is larger than or equal to a predetermined threshold value, the similarity measurement block 208 generates a next, e.g., second grid generation signal to the grid generation block 210; otherwise, it generates a completion signal to the grid generation block 210 and the switching block 212. The switching block 212 relays the deviation points and the displacements to the contour coder 24 in response to the completion signal from the similarity measurement block 208.
The grid generation block 210 selects, in response to the next grid generation signal, a grid of a next lower grid line density and provides grid information therefor to the feature point selection block 202; and the error detection process described above is repeated with respect to the newly selected grid. On the other hand, if the completion signal is received, the grid generation block 210 generates to the contour coder 24 a grid index indicating the grid which yields the similarity index smaller than the predetermined threshold value.
As demonstrated above, therefore, the present invention is capable of substantially reducing the volume of transmission data for representing the contour of regions contained in a video signal by transmitting the prediction parameters together with the deviation information including the deviation points and the displacements for the deviation points.
While the present invention has been described with respect to the particular embodiments, it will be apparent to those skilled in the art that various changes and modifications may be made without departing from the spirit and scope of the invention as defined in the following claims.

Claims (6)

What is claimed is:
1. An apparatus for encoding a contour of regions contained in a video signal, the video signal having a plurality of frames including a current frame and a previous frame, which comprises:
first segmentation means for segmenting the previous frame into a number of regions to generate a previous contour;
second segmentation means for segmenting the current frame into a number of regions to generate a current contour;
means for generating prediction parameters based on the previous and the current contours, and the previous and the current frames;
means for constructing a predicted current contour by transforming the previous contour based on the prediction parameters;
means for detecting deviation points lying on portions of the predicted current contour that deviate from the current contour, to thereby generate the deviation points and corresponding displacements to the deviation points, wherein each of the displacements represents a minimum distance in a horizontal direction between each of the deviation points and the current contour;
means for encoding the deviation points and the displacements together with the prediction parameters;
contour reconstruction means for relocating the deviation points based on the displacements and producing a reconstructed current contour from the predicted current contour;
means for taking N sample points on each of a plurality of line segments, each line segment joining two neighboring relocated deviation points on the reconstructed current contour, and calculating an error for each of the N sample points on each of the line segments to produce a set of errors for said each of the line segments, wherein the N sample points are equi-spaced on said each of the line segments and each of the errors represents a distance between each of the N sample points and the current contour;
means for summing up the set of errors to provide a similarity index;
control means for sequentially providing the feature point selection means with one or more grids until a similarity index obtained based on one of said one or more grids becomes smaller than a predetermined threshold value and generating a grid index indicating the grid related to the similarity index smaller than the predetermined threshold value; and
means for encoding the grid index;
wherein said prediction parameter generation means includes:
means for determining a predetermined number of vertex points on the previous contour to generate vertex information representing the positions of the vertex points on the previous contour;
means for generating vertex point blocks from the previous frame based on the vertex information, each of the vertex point blocks having a vertex point at the center thereof;
means for defining a search region of the current frame and generating a plurality of equal-sized candidate blocks from the search region, each of the candidate blocks is chosen such that its center pixel is a pixel forming a part of the current contour;
means for detecting predicted vertex points by performing a block matching algorithm between a vertex point block and each of the plurality of candidate blocks to generate predicted vertex information representing the positions of the predicted vertex points; and
means for generating the prediction parameters based on the vertex information and the predicted vertex information, and
wherein said deviation point detection means includes:
feature point selection means for selecting a number of points from the points on the predicted current contour as feature points by using a grid, wherein the feature points represent intersection points between the grid and the predicted current contour; and
means for determining the deviation points among the feature points, lying on portions of the predicted current contour that deviate from the current contour, and calculating the displacements for the deviation points, to thereby produce the deviation points and the corresponding displacements.
2. An apparatus for encoding a contour of regions contained in a video signal, the video signal having a plurality of frames including a current frame and a previous frame, which comprises:
first segmentation means for segmenting the previous frame into a number of regions to generate a previous contour;
second segmentation means for segmenting the current frame into a number of regions to generate a current contour;
means for generating prediction parameters based on the previous and the current contours, and the previous and the current frames;
means for constructing a predicted current contour by transforming the previous contour based on the prediction parameters;
a feature point selection means for selecting a number of points from the points on the predicted current contour as feature points by using a grid, wherein the feature points represent intersection points between the grid and the predicted current contour;
means for determining deviation points among the feature points, lying on portions of the predicted current contour that deviate from the current contour, and calculating displacements for the deviation points, wherein each of the displacements represents a minimum distance in a horizontal direction between each of the deviation points and the current contour;
means for relocating the deviation points based on the displacements and producing a reconstructed current contour from the predicted current contour;
means for taking N sample points on each of a plurality of line segments, each line segment joining two neighboring relocated deviation points on the reconstructed current contour, and calculating an error for each of the N sample points on each of the line segments to produce a set of errors for said each of the line segments, wherein the N sample points are equi-spaced on said each of the line segments and each of the errors represents a distance between each of the N sample points and the current contour;
means for summing up the set of errors to provide a similarity index;
control means for sequentially providing the feature point selection means with one or more grids until a similarity index obtained based on one of said one or more grids becomes smaller than a predetermined threshold value and generating a grid index indicating the grid related to the similarity index smaller than the predetermined threshold value; and
means for encoding the deviation points, the displacements, the prediction parameters, and the grid index.
3. The apparatus according to claim 2, wherein said prediction parameter generation means includes:
means for determining a predetermined number of vertex points on the previous contour to generate vertex information representing the positions of the vertex points on the previous contour;
means for generating vertex point blocks from the previous frame based on the vertex information, each of the vertex point blocks having a vertex point at the center thereof;
means for defining a search region of the current frame and generating a plurality of equal-sized candidate blocks from the search region, each of the candidate blocks is chosen such that its center pixel is a pixel forming a part of the current contour;
means for detecting predicted vertex points by performing a block matching algorithm between a vertex point block and each of the plurality of candidate blocks to generate predicted vertex information representing the positions of the predicted vertex points; and
means for generating the prediction parameters based on the vertex information and the predicted vertex information.
4. An apparatus for encoding a contour of regions contained in a video signal, the video signal having a plurality of frames including a current frame and a previous frame, which comprises:
first segmentation means for segmenting the previous frame into a number of regions to generate a previous contour;
second segmentation means for segmenting the current frame into a number of regions to generate a current contour;
means for generating prediction parameters based on the previous and the current contours, and the previous and the current frames;
means for constructing a predicted current contour by transforming the previous contour based on the prediction parameters;
means for detecting deviation points lying on portions of the predicted current contour that deviate from the current contour, to thereby generate the deviation points and corresponding displacements to the deviation points, wherein each of the displacements represents a minimum distance in a horizontal direction between each of the deviation points and the current contour;
means for encoding the deviation points and the displacements together with the prediction parameters;
contour reconstruction means for relocating the deviation points based on the displacements and producing a reconstructed current contour from the predicted current contour;
means for taking N sample points on each of a plurality of line segments, each line segment joining two neighboring relocated deviation points on the reconstructed current contour, and calculating an error for each of the N sample points on each of the line segments to produce a set of errors for said each of the line segments, wherein the N sample points are equi-spaced on said each of the line segments and each of the errors represents a distance between each of the N sample points and the current contour;
means for summing up the set of errors to provide a similarity index;
control means for sequentially providing the feature point selection means with one or more grids until a similarity index obtained based on one of said one or more grids becomes smaller than a predetermined threshold value and generating a grid index indicating the grid related to the similarity index smaller than the predetermined threshold value; and
means for encoding the grid index.
5. The apparatus according to claim 4, wherein said prediction parameter generation means includes:
means for determining a predetermined number of vertex points on the previous contour to generate vertex information representing the positions of the vertex points on the previous contour;
means for generating vertex point blocks from the previous frame based on the vertex information, each of the vertex point blocks having a vertex point at the center thereof;
means for defining a search region of the current frame and generating a plurality of equal-sized candidate blocks from the search region, each of the candidate blocks is chosen such that its center pixel is a pixel forming a part of the current contour;
means for detecting predicted vertex points by performing a block matching algorithm between a vertex point block and each of the plurality of candidate blocks to generate predicted vertex information representing the positions of the predicted vertex points; and
means for generating the prediction parameters based on the vertex information and the predicted vertex information.
6. The apparatus according to claim 4, wherein said deviation point detection means includes:
feature point selection means for selecting a number of points from the points on the predicted current contour as feature points by using a grid, wherein the feature points represent intersection points between the grid and the predicted current contour; and
means for determining the deviation points among the feature points, lying on portions of the predicted current contour that deviate from the current contour, and calculating the displacements for the deviation points, to thereby produce the deviation points and the corresponding displacements.
US08/580,237 1994-12-29 1995-12-28 Apparatus for encoding a contour of regions contained in a video signal Expired - Lifetime US5737449A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1019940038606A KR100235345B1 (en) 1994-12-29 1994-12-29 Moving picture estimation apparatus and method in divided region
KR94-38606 1994-12-29

Publications (1)

Publication Number Publication Date
US5737449A true US5737449A (en) 1998-04-07

Family

ID=19404832

Family Applications (1)

Application Number Title Priority Date Filing Date
US08/580,237 Expired - Lifetime US5737449A (en) 1994-12-29 1995-12-28 Apparatus for encoding a contour of regions contained in a video signal

Country Status (4)

Country Link
US (1) US5737449A (en)
JP (1) JPH08275180A (en)
KR (1) KR100235345B1 (en)
CN (1) CN1130919C (en)

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5828790A (en) * 1995-12-23 1998-10-27 Daewoo Electronics Co., Ltd. Method and apparatus for approximating a contour image of an object in a video signal
WO1999005851A2 (en) * 1997-07-28 1999-02-04 Physical Optics Corporation Method of isomorphic singular manifold projection still/video imagery compression
US5881175A (en) * 1996-06-07 1999-03-09 Daewoo Electronics Co., Ltd. Method and apparatus for encoding an image signal by using the contour signal thereof
US5881174A (en) * 1997-02-18 1999-03-09 Daewoo Electronics Co., Ltd. Method and apparatus for adaptively coding a contour of an object
US5883977A (en) * 1996-12-30 1999-03-16 Daewoo Electronics Co., Ltd. Method and apparatus for encoding a video signal of a contour of an object
US5896467A (en) * 1997-01-10 1999-04-20 Daewoo Electronics Co., Ltd. Method and apparatus for encoding a contour image of an object in a video signal
US5896466A (en) * 1997-01-20 1999-04-20 Daewoo Electronics Co., Ltd. Method and apparatus for encoding a contour of an object by using a vertex inserting technique
US5912991A (en) * 1997-02-07 1999-06-15 Samsung Electronics Co., Ltd. Contour encoding method using error bands
US5915045A (en) * 1997-01-15 1999-06-22 Daewoo Electronics Co., Ltd. Method and apparatus for encoding a contour of an object by adapting a vertex coding technique
US5929917A (en) * 1997-04-11 1999-07-27 Daewoo Electronics Co., Ltd. Method and apparatus for adaptively coding a contour of an object
US5978510A (en) * 1997-03-20 1999-11-02 Hyundai Electronics Industries Co., Ltd. Method and apparatus for predictively coding contour of video signal
US5995671A (en) * 1996-11-08 1999-11-30 Hughes Electronics Corporation Efficient representation of texture for gray scale alpha maps
US6011588A (en) * 1997-04-11 2000-01-04 Daewoo Electronics Co., Ltd. Method and apparatus for coding a contour of an object employing temporal correlation thereof
US6055337A (en) * 1996-07-09 2000-04-25 Daewoo Electronics Co. Ltd. Method and apparatus for encoding a contour of an object expressed in a video signal
US6104404A (en) * 1994-09-27 2000-08-15 International Business Machines Corporation Drawing candidate line segments extraction system and method and solid model synthesis system and method
US6121980A (en) * 1997-01-10 2000-09-19 Samsung Electronics Co., Ltd. Vertex-based predictive shape coding method using adaptive sampling direction selection
US6128041A (en) * 1997-07-11 2000-10-03 Daewoo Electronics Co., Ltd. Method and apparatus for binary shape encoding
US20010025298A1 (en) * 2000-03-17 2001-09-27 Koichi Masukura Object region data generating method, object region data generating apparatus, approximation polygon generating method, and approximation polygon generating apparatus
US20020025074A1 (en) * 2000-08-28 2002-02-28 Wilhelm Robert G. Systems and methods for adaptive sampling and estimating a systematic relationship between a plurality of points
US20030012278A1 (en) * 2001-07-10 2003-01-16 Ashish Banerji System and methodology for video compression
US20040266015A1 (en) * 2002-12-20 2004-12-30 Dakocytomation Denmark A/S Automated sample processing apparatus and a method of automated treating of samples and use of such apparatus
US6944351B2 (en) * 1995-10-26 2005-09-13 Hyundai Curitel, Inc. Apparatus and method of encoding/decoding a coded block pattern
US20050226329A1 (en) * 2003-05-20 2005-10-13 Do-Yeon Weon Motion estimation method using multilevel succesive elimination algorithm
US20060034527A1 (en) * 2003-04-21 2006-02-16 Aleksandr Gnitsevich Method for encoding co-ordinates of a video image moving along the display of a computing device
US20060262845A1 (en) * 1999-04-17 2006-11-23 Adityo Prakash Segment-based encoding system using segment hierarchies
US20070183663A1 (en) * 2006-02-07 2007-08-09 Haohong Wang Intra-mode region-of-interest video object segmentation
US20070183661A1 (en) * 2006-02-07 2007-08-09 El-Maleh Khaled H Multi-mode region-of-interest video object segmentation
US20070183662A1 (en) * 2006-02-07 2007-08-09 Haohong Wang Inter-mode region-of-interest video object segmentation
WO2009113754A1 (en) * 2008-03-12 2009-09-17 Atlab Inc. Touch sensor device and pointing coordinate determination method thereof
US20090316998A1 (en) * 2008-06-20 2009-12-24 Konica Minolta Business Technologies, Inc. Outlining method for properly representing curved line and straight line, and image compression method using the same
CN102270342A (en) * 2010-06-01 2011-12-07 西门子公司 Detecting segmentation method for a structure in image data
US20140105473A1 (en) * 2012-10-12 2014-04-17 Samsung Electronics Co., Ltd. Method and apparatus for segmenting contour in image
US20210227261A1 (en) * 2016-02-01 2021-07-22 Lg Electronics Inc. Method and apparatus for encoding/decoding video signal by using edge-adaptive graph-based transform

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4034380B2 (en) 1996-10-31 2008-01-16 株式会社東芝 Image encoding / decoding method and apparatus
KR101681303B1 (en) * 2010-07-29 2016-12-01 에스케이 텔레콤주식회사 Method and Apparatus for Encoding/Decoding of Video Data Using Partitioned-Block Prediction

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3632865A (en) * 1969-12-23 1972-01-04 Bell Telephone Labor Inc Predictive video encoding using measured subject velocity
US4703350A (en) * 1985-06-03 1987-10-27 Picturetel Corporation Method and apparatus for efficiently communicating image sequences
US5537155A (en) * 1994-04-29 1996-07-16 Motorola, Inc. Method for estimating motion in a video sequence
US5557341A (en) * 1991-04-12 1996-09-17 Dv Sweden Ab Iterative method for estimating motion content in video signals using successively reduced block size
US5594813A (en) * 1992-02-19 1997-01-14 Integrated Information Technology, Inc. Programmable architecture and methods for motion estimation
US5598215A (en) * 1993-05-21 1997-01-28 Nippon Telegraph And Telephone Corporation Moving image encoder and decoder using contour extraction
US5608458A (en) * 1994-10-13 1997-03-04 Lucent Technologies Inc. Method and apparatus for a region-based approach to coding a sequence of video images

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE3933346C1 (en) * 1989-10-06 1991-04-04 Ant Nachrichtentechnik Gmbh, 7150 Backnang, De

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3632865A (en) * 1969-12-23 1972-01-04 Bell Telephone Labor Inc Predictive video encoding using measured subject velocity
US4703350A (en) * 1985-06-03 1987-10-27 Picturetel Corporation Method and apparatus for efficiently communicating image sequences
US5557341A (en) * 1991-04-12 1996-09-17 Dv Sweden Ab Iterative method for estimating motion content in video signals using successively reduced block size
US5594813A (en) * 1992-02-19 1997-01-14 Integrated Information Technology, Inc. Programmable architecture and methods for motion estimation
US5598215A (en) * 1993-05-21 1997-01-28 Nippon Telegraph And Telephone Corporation Moving image encoder and decoder using contour extraction
US5537155A (en) * 1994-04-29 1996-07-16 Motorola, Inc. Method for estimating motion in a video sequence
US5608458A (en) * 1994-10-13 1997-03-04 Lucent Technologies Inc. Method and apparatus for a region-based approach to coding a sequence of video images

Cited By (64)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6104404A (en) * 1994-09-27 2000-08-15 International Business Machines Corporation Drawing candidate line segments extraction system and method and solid model synthesis system and method
US6944351B2 (en) * 1995-10-26 2005-09-13 Hyundai Curitel, Inc. Apparatus and method of encoding/decoding a coded block pattern
US5828790A (en) * 1995-12-23 1998-10-27 Daewoo Electronics Co., Ltd. Method and apparatus for approximating a contour image of an object in a video signal
US5881175A (en) * 1996-06-07 1999-03-09 Daewoo Electronics Co., Ltd. Method and apparatus for encoding an image signal by using the contour signal thereof
US6055337A (en) * 1996-07-09 2000-04-25 Daewoo Electronics Co. Ltd. Method and apparatus for encoding a contour of an object expressed in a video signal
US5995671A (en) * 1996-11-08 1999-11-30 Hughes Electronics Corporation Efficient representation of texture for gray scale alpha maps
US5883977A (en) * 1996-12-30 1999-03-16 Daewoo Electronics Co., Ltd. Method and apparatus for encoding a video signal of a contour of an object
US5896467A (en) * 1997-01-10 1999-04-20 Daewoo Electronics Co., Ltd. Method and apparatus for encoding a contour image of an object in a video signal
US6121980A (en) * 1997-01-10 2000-09-19 Samsung Electronics Co., Ltd. Vertex-based predictive shape coding method using adaptive sampling direction selection
US5915045A (en) * 1997-01-15 1999-06-22 Daewoo Electronics Co., Ltd. Method and apparatus for encoding a contour of an object by adapting a vertex coding technique
US5896466A (en) * 1997-01-20 1999-04-20 Daewoo Electronics Co., Ltd. Method and apparatus for encoding a contour of an object by using a vertex inserting technique
US5912991A (en) * 1997-02-07 1999-06-15 Samsung Electronics Co., Ltd. Contour encoding method using error bands
US5881174A (en) * 1997-02-18 1999-03-09 Daewoo Electronics Co., Ltd. Method and apparatus for adaptively coding a contour of an object
US5978510A (en) * 1997-03-20 1999-11-02 Hyundai Electronics Industries Co., Ltd. Method and apparatus for predictively coding contour of video signal
US6011588A (en) * 1997-04-11 2000-01-04 Daewoo Electronics Co., Ltd. Method and apparatus for coding a contour of an object employing temporal correlation thereof
US5929917A (en) * 1997-04-11 1999-07-27 Daewoo Electronics Co., Ltd. Method and apparatus for adaptively coding a contour of an object
US6128041A (en) * 1997-07-11 2000-10-03 Daewoo Electronics Co., Ltd. Method and apparatus for binary shape encoding
WO1999005851A3 (en) * 1997-07-28 1999-04-15 Physical Optics Corp Method of isomorphic singular manifold projection still/video imagery compression
WO1999005851A2 (en) * 1997-07-28 1999-02-04 Physical Optics Corporation Method of isomorphic singular manifold projection still/video imagery compression
US20060262845A1 (en) * 1999-04-17 2006-11-23 Adityo Prakash Segment-based encoding system using segment hierarchies
US20010025298A1 (en) * 2000-03-17 2001-09-27 Koichi Masukura Object region data generating method, object region data generating apparatus, approximation polygon generating method, and approximation polygon generating apparatus
US20020025074A1 (en) * 2000-08-28 2002-02-28 Wilhelm Robert G. Systems and methods for adaptive sampling and estimating a systematic relationship between a plurality of points
US7787696B2 (en) * 2000-08-28 2010-08-31 University Of North Carolina At Charlotte Systems and methods for adaptive sampling and estimating a systematic relationship between a plurality of points
US20030012278A1 (en) * 2001-07-10 2003-01-16 Ashish Banerji System and methodology for video compression
US9894379B2 (en) * 2001-07-10 2018-02-13 The Directv Group, Inc. System and methodology for video compression
US7758809B2 (en) 2002-12-20 2010-07-20 Dako Cytomation Denmark A/S Method and system for pretreatment of tissue slides
US8394635B2 (en) 2002-12-20 2013-03-12 Dako Denmark A/S Enhanced scheduling sample processing system and methods of biological slide processing
US8788217B2 (en) 2002-12-20 2014-07-22 Dako Denmark A/S Information notification sample processing system and methods of biological slide processing
US8969086B2 (en) 2002-12-20 2015-03-03 Dako Denmark A/S Enhanced scheduling sample processing system and methods of biological slide processing
US9229016B2 (en) 2002-12-20 2016-01-05 Dako Denmark A/S Information notification sample processing system and methods of biological slide processing
US10156580B2 (en) 2002-12-20 2018-12-18 Dako Denmark A/S Information notification sample processing system and methods of biological slide processing
US20040266015A1 (en) * 2002-12-20 2004-12-30 Dakocytomation Denmark A/S Automated sample processing apparatus and a method of automated treating of samples and use of such apparatus
US7648678B2 (en) 2002-12-20 2010-01-19 Dako Denmark A/S Method and system for pretreatment of tissue slides
US8673642B2 (en) 2002-12-20 2014-03-18 Dako Denmark A/S Enhanced scheduling sample processing system and methods of biological slide processing
US8663978B2 (en) 2002-12-20 2014-03-04 Dako Denmark A/S Method and apparatus for automatic staining of tissue samples
US20060063265A1 (en) * 2002-12-20 2006-03-23 Dakocytomation Denmark A/S Advance programmed sample processing system and methods of biological slide processing
US7937228B2 (en) 2002-12-20 2011-05-03 Dako Denmark A/S Information notification sample processing system and methods of biological slide processing
US7960178B2 (en) 2002-12-20 2011-06-14 Dako Denmark A/S Enhanced scheduling sample processing system and methods of biological slide processing
US9778273B2 (en) 2002-12-20 2017-10-03 Dako Denmark A/S Isolated communication sample processing system and methods of biological slide processing
US8529836B2 (en) 2002-12-20 2013-09-10 Dako Denmark A/S Apparatus for automated processing biological samples
US9599630B2 (en) 2002-12-20 2017-03-21 Dako Denmark A/S Method and apparatus for automatic staining of tissue samples
US8216512B2 (en) 2002-12-20 2012-07-10 Dako Denmark A/S Apparatus for automated processing biological samples
US8257968B2 (en) 2002-12-20 2012-09-04 Dako Denmark A/S Method and apparatus for automatic staining of tissue samples
US8784735B2 (en) 2002-12-20 2014-07-22 Dako Denmark A/S Apparatus for automated processing biological samples
US8386195B2 (en) 2002-12-20 2013-02-26 Dako Denmark A/S Information notification sample processing system and methods of biological slide processing
US8298815B2 (en) 2002-12-20 2012-10-30 Dako Denmark A/S Systems and methods of sample processing and temperature control
US20060034527A1 (en) * 2003-04-21 2006-02-16 Aleksandr Gnitsevich Method for encoding co-ordinates of a video image moving along the display of a computing device
US20050226329A1 (en) * 2003-05-20 2005-10-13 Do-Yeon Weon Motion estimation method using multilevel succesive elimination algorithm
US8265349B2 (en) 2006-02-07 2012-09-11 Qualcomm Incorporated Intra-mode region-of-interest video object segmentation
US8265392B2 (en) * 2006-02-07 2012-09-11 Qualcomm Incorporated Inter-mode region-of-interest video object segmentation
US8150155B2 (en) 2006-02-07 2012-04-03 Qualcomm Incorporated Multi-mode region-of-interest video object segmentation
US8605945B2 (en) 2006-02-07 2013-12-10 Qualcomm, Incorporated Multi-mode region-of-interest video object segmentation
US20070183662A1 (en) * 2006-02-07 2007-08-09 Haohong Wang Inter-mode region-of-interest video object segmentation
US20070183661A1 (en) * 2006-02-07 2007-08-09 El-Maleh Khaled H Multi-mode region-of-interest video object segmentation
US20070183663A1 (en) * 2006-02-07 2007-08-09 Haohong Wang Intra-mode region-of-interest video object segmentation
US20110037727A1 (en) * 2008-03-12 2011-02-17 Atlab Inc. Touch sensor device and pointing coordinate determination method thereof
WO2009113754A1 (en) * 2008-03-12 2009-09-17 Atlab Inc. Touch sensor device and pointing coordinate determination method thereof
US20090316998A1 (en) * 2008-06-20 2009-12-24 Konica Minolta Business Technologies, Inc. Outlining method for properly representing curved line and straight line, and image compression method using the same
US8682074B2 (en) * 2010-06-01 2014-03-25 Siemens Aktiengesellschaft Method for checking the segmentation of a structure in image data
US20120134552A1 (en) * 2010-06-01 2012-05-31 Thomas Boettger Method for checking the segmentation of a structure in image data
CN102270342A (en) * 2010-06-01 2011-12-07 西门子公司 Detecting segmentation method for a structure in image data
US20140105473A1 (en) * 2012-10-12 2014-04-17 Samsung Electronics Co., Ltd. Method and apparatus for segmenting contour in image
US20210227261A1 (en) * 2016-02-01 2021-07-22 Lg Electronics Inc. Method and apparatus for encoding/decoding video signal by using edge-adaptive graph-based transform
US11695958B2 (en) * 2016-02-01 2023-07-04 Lg Electronics Inc. Method and apparatus for encoding/decoding video signal by using edge-adaptive graph-based transform

Also Published As

Publication number Publication date
CN1130847A (en) 1996-09-11
JPH08275180A (en) 1996-10-18
KR960028547A (en) 1996-07-22
KR100235345B1 (en) 1999-12-15
CN1130919C (en) 2003-12-10

Similar Documents

Publication Publication Date Title
US5737449A (en) Apparatus for encoding a contour of regions contained in a video signal
EP0801504B1 (en) Method for encoding a contour of an object in a video signal by using a contour motion estimation technique
US5686973A (en) Method for detecting motion vectors for use in a segmentation-based coding system
US5978512A (en) Polygonal approximation method and apparatus for use in a contour encoding system
US5995668A (en) Segmented picture coding method and system, and corresponding decoding method and system
US5774595A (en) Contour approximation method for representing a contour of an object
US5691769A (en) Apparatus for encoding a contour of an object
US5805221A (en) Video signal coding system employing segmentation technique
JP3917691B2 (en) Outline coding method and apparatus
EP0889652A2 (en) Method and apparatus for encoding a contour of an object based on a contour motion estimation technique
EP0734163B1 (en) A contour approximation apparatus for representing a contour of an object
US5774596A (en) Adaptive contour coding method for encoding a contour image in a video signal
EP0871331B1 (en) Method and apparatus for adaptively coding a contour of an object
EP0734168A2 (en) Apparatus for encoding a moving image signal having a still object
US5896467A (en) Method and apparatus for encoding a contour image of an object in a video signal
US5828790A (en) Method and apparatus for approximating a contour image of an object in a video signal
US6163627A (en) Method and apparatus for encoding a contour of an object
JPH10507331A (en) Method and system for encoding an image sequence and associating the encoded signal with a storage medium, and method and system for decoding such an encoded signal
US5754703A (en) Method for encoding a contour of an object in a video signal
JP3694349B2 (en) Contour coding device
Ozcelik et al. Detection and encoding of model failures in very low bit rate video coding
GB2321359A (en) Polygonal approximation in video contour encoding system

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

AS Assignment

Owner name: DAEWOO ELECTRONICS CORPORATION, KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DAEWOO ELECTRONICS CO., LTD.;REEL/FRAME:013645/0159

Effective date: 20021231

FPAY Fee payment

Year of fee payment: 8

FPAY Fee payment

Year of fee payment: 12

AS Assignment

Owner name: MAPLE VISION TECHNOLOGIES INC., CANADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DAEWOO ELECTRONICS CORPORATION;REEL/FRAME:027437/0345

Effective date: 20111215

AS Assignment

Owner name: QUARTERHILL INC., CANADA

Free format text: MERGER AND CHANGE OF NAME;ASSIGNORS:MAPLE VISION TECHNOLOGIES INC.;QUARTERHILL INC.;REEL/FRAME:042935/0282

Effective date: 20170601

AS Assignment

Owner name: WI-LAN INC., CANADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:QUARTERHILL INC.;REEL/FRAME:043182/0859

Effective date: 20170601