WO1999003068A1 - Method and apparatus for monoscopic to stereoscopic image conversion - Google Patents

Method and apparatus for monoscopic to stereoscopic image conversion Download PDF

Info

Publication number
WO1999003068A1
WO1999003068A1 PCT/US1998/012119 US9812119W WO9903068A1 WO 1999003068 A1 WO1999003068 A1 WO 1999003068A1 US 9812119 W US9812119 W US 9812119W WO 9903068 A1 WO9903068 A1 WO 9903068A1
Authority
WO
WIPO (PCT)
Prior art keywords
perspective
sequence
image
frames
outline
Prior art date
Application number
PCT/US1998/012119
Other languages
French (fr)
Inventor
Kelvin Ma
David Swift
Original Assignee
Reveo, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Reveo, Inc. filed Critical Reveo, Inc.
Priority to AU80668/98A priority Critical patent/AU8066898A/en
Publication of WO1999003068A1 publication Critical patent/WO1999003068A1/en

Links

Classifications

    • G06T3/18
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/261Image signal generators with monoscopic-to-stereoscopic image conversion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/122Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/189Recording image signals; Reproducing recorded image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/221Image signal generators using stereoscopic image cameras using a single 2D image sensor using the relative movement between cameras and objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/257Colour aspects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0092Image segmentation from stereoscopic image signals

Definitions

  • This invention relates to a method and an apparatus for converting monoscopic (i.e., two- dimensional) images into stereoscopic (i.e., three- dimensional) images, and more particularly using image warping.
  • most images are recorded in monoscopic format on recording medium (e.g., motion picture film and video media) where some depth information is lost.
  • a monoscopic camera lens produces an image according to the laws of Cartesian space and perspective. If such an image could be processed to extract that information, and thus be converted into three-dimensional stereoscopic format, there could be a vast amount of stereoscopic material available for three-dimensional viewing.
  • the horizontal shift method isolates several important objects in the image field and shifts them horizontally.
  • B.J. Garcia, Ph.D. and R.J. Greenstein, M.D. "True- Stereoscopic Video Media from Monoscopic Sources: The DeepVision System for Minimally Invasive Surgery, " Medi cal Applica ti ons , Virtual Reality Systems, pp . 52-57 (1 994) .
  • the amount of the shift for each object varies according to its depth arrangement.
  • the depth arrangement also referred to as stereoscopic depth cue or binocular disparity
  • This process creates a stereoscopic image with depth perception between objects, but the objects themselves lack depth and appear to be flat cardboard cutouts .
  • Another method uses moving images to extract a second perspective.
  • H. Murata, T. Okino, T. Iinuma, S. Yamashita, S. Tanase, K. Terada and K. Kanatani "Conversion of Two-Dimensional Images to Three Dimensions," SID 95 Digest , Society for Information Display 1995 Symposium Proceedings (1995) .
  • a series of different perspectives is generated due to either the object's movement or the camera's motion. Different perspectives from different images in a time series can then be extracted to generate a second perspective for the stereoscopic effect.
  • Possible applications for such a method and apparatus include rapid conversion of existing monoscopic movie and video media, as well as still photograph (i.e., single image), libraries to stereoscopic images.
  • FIG. 1 illustrates a system for converting monoscopic images to stereoscopic images
  • FIG. 2 is an actual perspective calculation curve
  • FIG. 3 illustrates a second perspective image geometry
  • FIG. 4 illustrates a sequence of successive image frames and the image frame horizontal axis
  • FIGS. 5a and 5b show pixel sampling
  • FIG. 6 is a modified (saturating) perspective calculation curve
  • FIGS. 7a and 7b illustrate the transformation of a cube
  • FIG. 8 illustrates vector clipping
  • FIG. 9a is a flow diagram of a preferred embodiment of the method implemented by the apparatus.
  • FIG. 9b is a flow diagram of the tweening process used in the method shown in FIG. 9a.
  • the Appendix contains a listing of the symbolic mathematical program used to derive a generalized equation for shift.
  • the present invention is a method and an apparatus for converting monoscopic images to stereoscopic images .
  • Stereoscopic images require two slightly offset perspective images in order to create a stereoscopic presentation with the appearance of depth.
  • Most existing images stored in motion picture film, photographic or video media format contain only a single perspective and, therefore, appear flat.
  • a second perspective image must be obtained.
  • the second perspective image is preferably only slightly offset from the first perspective image.
  • a system of techniques of shifting and warping objects within the first image can be implemented to automate the generation of the second perspective image.
  • the technique of the present invention for converting monoscopic into stereoscopic images uses image warping techniques often used for morphmg image objects from one shape into a different shape. Warping comprises two steps.
  • the object's outlined shape is transformed into a new shape.
  • This transformation preferably is primarily done by pixel interpolation.
  • the interpolation can be either linear (for speed) or quadratic (for accuracy) .
  • the pixels within the outlined shape of the image are interpolated where the boundary of the image outline expands or contracts.
  • pixel interpolation horizontal to the image frame axis (indicated by x in FIG. 4) is required, since the second perspective image is a slight horizontal shift from the first perspective image. This simplifies the procedure of image warping and helps speed up the transformation calculations. Indeed, a simple horizontal linear pixel interpolation can be achieved in close to real-time.
  • Objects are the components of an image which are manipulated to give the image depth. Examples of objects are tables, chairs, walls, floors, lights, people, etc. Objects can further be divided into sub-objects to add more depth. For example, a person could be broken into two arm objects, a head object, a body object, and two leg objects.
  • a frame is a single image at a specific time.
  • NTSC video consists of 30 frames per second or 30 images per second. Sequential frames are the sequential images to be converted.
  • a key frame is a specific frame or image where input from the user is required to help direct the animation.
  • the start and end of a new scene are key frames, ._.Frames where objects start or stop motion are typically key frames.
  • Clipping occurs when one object (the foreground object) obscures another object (the background object) .
  • the portions of the background object which are obscured are removed or not displayed and the foreground object is displayed instead.
  • a "tweening" process preferably is implemented to automate the outlining process.
  • the tweening process involves interpolations of the vector outlined shape of all the outlined objects within the scene from one key frame ("the start key frame") to the next key frame (“the stop key frame”).
  • the interpolation can be either linear or nonlinear (e.g., spline) depending on accuracy and/or speed requirements. Only the end points (i.e., the key frames) of the motion path for the objects need to be defined, significantly reducing the burden on the operator.
  • a first set of outlines is provided either by an automatic image recognition process and/or manually by an operator to outline vectors on all the necessary key objects within an image. This set of outlines is then shifted and warped through the perspective transformation based on the depth arrangement of the objects provided by the operator.
  • objects are outlined, a set of vectors are used to represent this.
  • vectors (outlines) from two objects intersect, vector-crossovers are created.
  • the background vectors should be clipped to conform to the foreground vectors when this happens.
  • Vector cross-overs can result in undesirable effects, therefore, a second set of outlined objects is generated after the perspective transformation.
  • FIG. 1 shows two-dimensional video media images comprism-g unprocessed still frames 2 provided to workstation 3 through digital image capture using video media frame grabber 1. Processed still frames 4 are displayed in sequence on monitor 5.
  • Key objects are operator-selected foreground objects in a particular sequence of images.
  • Key frames which correspond to frames m which key objects change motion or deviate from a predicted motion path, can be identified automatically by delta pixel changes between frames. Delta pixel changes can also be used to determine the basic motion path within each scene.
  • the key frames can be automatically selected based on the extreme points of the motion path.
  • the key frames can also be selected by a human operator after viewing the scene. These key frames are then individually processed by outlining the key objects within the image.
  • Objects can be detected automatically by an edge detection algorithm or manually by an operator.
  • Automatic edge detection algorithms such as Laplacian, Sobel and Prewitt convoluted filters can be used to highlight the object's outline. Such algorithms are described, for example, in Christopher Watkms, Alberto Sadun, Stephen Marenka, "Modern Image Processing: Warping, Morphing, and Classical Techniques," Academic Press Professional, Inc. (1993) .
  • the computer can then trace around the highlighted objects to form vector outlines.
  • a vector is a line which connects points (i.e., the control points) on the object.
  • Key objects can be identified by the operator as corresponding to certain predefined objects for predefined profile warping.
  • Each predefined object preferably has a profile identification which defines its perceived surface profile. This predefined profile warping eliminates the need to define vectors within an object in order to recreate its full volume appearance.
  • the object's outline defines the area which will be transformed via perspective warping.
  • Perspective warping is a type of image warping in which the image is horizontally shifted. Therefore, the actual transformation is significantly faster to compute because there is no vertical warping.
  • the object's outline is warped and the object's image within the outlined area is then warped into the necessary perspective.
  • the control points of the outline are transformed into their new shape and position based on the depth arrangements and/or the object's profile identification.
  • the actual transformation only shifts the end points of the vectors horizontally.
  • the amount of the shift may be based, for example, on the computed depth as shown m FIG. 2, which shows the relationship between the distance from the eyes, y, versus the amount of relative shift, ⁇ , at any horizontal point x for a fixed focus distance of 45 mcnes and an eye separation distance of 3 inches.
  • this position offset distance saturates rapidly as the depth becomes farther away.
  • the neutral plane (also referred to as the "zero- shifted" plane) is that plane on which no point shifts between the left and right perspective image. Any point on a plane m front of the neutral plane shifts to the left and any point on a plane behind the neutral plane shifts to the right.
  • the selection of the neutral plane is important to minimize the maximum shift required by the object.
  • the neutral plane can be selected automatically by choosing the center plane between the front-most object and the back-most object. Manual adjustment by the operator is allowed for aesthetic reasons if the operator so chooses.
  • Objects in front of the neutral plane appear to float out of the viewing screen. This can produce undesirable effects due to object's partial appearance at the border of the screen. This creates an unnatural appearance as the object gets partially clipped by the border behind it instead of m front. This can be corrected by shifting the left and right frames horizontally so that the object is behind the viewing screen.
  • FIGS. 7a and 7b show the transformation of a simple cube.
  • FIG. 7a shows the left perspective image.
  • FIG. 7b shows the right perspective image which results from the transformation.
  • the main outlines are the edges of the cube and the control points are the eight corners of the cube.
  • the control points are manipulated and depth arranged by an operator to determine the current position in depth. This perceived depth is then horizontally translated right to create the second perspective image.
  • the resulting stereoscopic image can be displayed in quasi-real-time on a monitor to provide feedback to the operator.
  • image warping based on the vector outline transformation is performed based on the left perspective image.
  • a new set of vectors is generated for use in the warping procedure based on the object's depth arrangement. Any vector outlined object in the background which crosses vectors with the foreground objects will be clipped around the contour of the foreground object to prevent vector crossover during the warping procedure, as shown in FIG. 8. Certain vectors in the new set of vectors are allowed to crossover when the distance of the crossover is sufficiently small, thereby allowing the foreground image to fold over the background image .
  • the frame rate When motion picture film is converted to video media, the frame rate must be converted from 24 frames per second to 30 frames per second. One common way to do this is by making every fifth video media frame the duplicate of the preceding motion picture film frame. As a way of reducing the number of frames to be processed and increasing the interpolation accuracy, such redundant frames may be removed from video media prior to monoscopic to stereoscopic conversion. Of course, the redundancy must be remtroduced to the video media after processing is complete.
  • step 30 If the digital images are from video media and were originally recorded on motion picture film, redundant frames are removed at step 30. Key objects are selected at step 40. Key frames are identified at step 50. Each frame between the start key frame and the stop key frame is tweened at step 60. Finally, the sequence of stereoscopic images is displayed at step 70.
  • the tweening process of step 60 begins with outlining the objects to form vector outlines at step 100. The depth of objects is set at step 110. If a vector outlined object in the background crosses vectors with a foreground object, the vectors of the background object are clipped at step 130. If the object corresponds to a predefined object it is transformed using a predefined profile warping at step 150. Otherwise, the object is transformed using a selected perspective calculation curve at step 160. Unless a stop key frame is reached, process returns to step 100. The tweening process is repeated for each set of start and stop key frames.

Abstract

A method and apparatus for converting a monoscopic image with one perspective to a stereoscopic image with two perspectives is provided. A foreground object is identified and outlined. The depth of the object is endowed with depth information. The second perspective is generated by image warping. Alternate embodiments comprising additional features, such as removing redundant frames, which reduce the necessary computation are also provided.

Description

METHOD AND APPARATUS FOR MONOSCOPIC TO STEREOSCOPIC IMAGE CONVERSION
Background of the Invention
This invention relates to a method and an apparatus for converting monoscopic (i.e., two- dimensional) images into stereoscopic (i.e., three- dimensional) images, and more particularly using image warping.
The human brain perceives depth by receiving two different images, which are horizontally shifted relative to each other, from each eye. Such depth information is important for determining how far away an object is and the depth arrangements between various objects in a scene. Unfortunately, most images are recorded in monoscopic format on recording medium (e.g., motion picture film and video media) where some depth information is lost. However, even a monoscopic camera lens produces an image according to the laws of Cartesian space and perspective. If such an image could be processed to extract that information, and thus be converted into three-dimensional stereoscopic format, there could be a vast amount of stereoscopic material available for three-dimensional viewing. There are several methods to generate stereoscopic images from monoscopic images. One method, the horizontal shift method, isolates several important objects in the image field and shifts them horizontally. B.J. Garcia, Ph.D. and R.J. Greenstein, M.D., "True- Stereoscopic Video Media from Monoscopic Sources: The DeepVision System for Minimally Invasive Surgery, " Medi cal Applica ti ons , Virtual Reality Systems, pp . 52-57 (1 994) . The amount of the shift for each object varies according to its depth arrangement. The depth arrangement (also referred to as stereoscopic depth cue or binocular disparity) of an object can be set manually using cues such as shadows and the perspective within the monoscopic image. This process creates a stereoscopic image with depth perception between objects, but the objects themselves lack depth and appear to be flat cardboard cutouts .
Another method, the frame delay method, uses moving images to extract a second perspective. H. Murata, T. Okino, T. Iinuma, S. Yamashita, S. Tanase, K. Terada and K. Kanatani, "Conversion of Two-Dimensional Images to Three Dimensions," SID 95 Digest , Society for Information Display 1995 Symposium Proceedings (1995) . When an object is in motion, a series of different perspectives is generated due to either the object's movement or the camera's motion. Different perspectives from different images in a time series can then be extracted to generate a second perspective for the stereoscopic effect. Though this technique can create very realistic three-dimensional effects, it requires that motion of both the camera and the objects be limited to a direction horizontal to the image frame's horizontal axis. The camera must move in a horizontal fashion perpendicular to the subject matter (this is called a "truck" in the motion picture industry) . The reason for this perpendicular motion is that each stereoscopic image pair is formed by pairing up images which are delayed from each other. Because the camera is limited to sideways motion, the delay between images results in a parallel camera separation. If the camera were not moving in a perpendicular and horizontal manner, the image pairs would not be parallel and would not form a good stereoscopic image. Also, since this technique requires a series of images taken at different times, a single image cannot be converted.
In view of the foregoing, it would be desirable to be able to provide a method and an apparatus for converting monoscopic images into stereoscopic images which gives the appearance of depth to objects themselves, as well as the appearance of depth between objects. It would also be desirable to be able to convert images with vertical as well as horizontal object motion relative to the frame's horizontal axis. It would also be desirable to be able to convert a single image.
Summary of the Invention
It is an object of the present invention to provide a method and an apparatus for converting monoscopic images into stereoscopic images which gives the appearance of depth to objects themselves, as well as the appearance of depth between objects.
It is also an object of the present invention to provide a method and an apparatus for converting images with vertical as well as horizontal object motion relative to the- frame's horizontal axis.
It is another object of the present invention to provide a method and an apparatus for converting a single image .
These and other objects of the invention are accomplished in accordance with the principles of the invention by identifying at least one object in the monoscopic image with a first perspective, outlining the object to generate an outline, endowing the object with depth information and generating a second perspective by image warping the object.
Possible applications for such a method and apparatus include rapid conversion of existing monoscopic movie and video media, as well as still photograph (i.e., single image), libraries to stereoscopic images.
Brief Description of the Drawings
The above and other objects and advantages of the invention will be apparent upon consideration of the following detailed description, taken in conjunction with the accompanying drawings, in which the reference characters refer to like parts throughout and in which:
FIG. 1 illustrates a system for converting monoscopic images to stereoscopic images; FIG. 2 is an actual perspective calculation curve;
FIG. 3 illustrates a second perspective image geometry;
FIG. 4 illustrates a sequence of successive image frames and the image frame horizontal axis; FIGS. 5a and 5b show pixel sampling; FIG. 6 is a modified (saturating) perspective calculation curve;
FIGS. 7a and 7b illustrate the transformation of a cube; FIG. 8 illustrates vector clipping;
FIG. 9a is a flow diagram of a preferred embodiment of the method implemented by the apparatus; and
FIG. 9b is a flow diagram of the tweening process used in the method shown in FIG. 9a. The Appendix contains a listing of the symbolic mathematical program used to derive a generalized equation for shift.
Detailed Description of the Invention
The present invention is a method and an apparatus for converting monoscopic images to stereoscopic images .
Stereoscopic images require two slightly offset perspective images in order to create a stereoscopic presentation with the appearance of depth. Most existing images stored in motion picture film, photographic or video media format contain only a single perspective and, therefore, appear flat. In order to create a three- dimensional image, a second perspective image must be obtained. The second perspective image is preferably only slightly offset from the first perspective image. In accordance with the present invention, it is possible to generate the second perspective image from the first image. A system of techniques of shifting and warping objects within the first image can be implemented to automate the generation of the second perspective image. The technique of the present invention for converting monoscopic into stereoscopic images uses image warping techniques often used for morphmg image objects from one shape into a different shape. Warping comprises two steps. First, the object's outlined shape is transformed into a new shape. This transformation preferably is primarily done by pixel interpolation. The interpolation can be either linear (for speed) or quadratic (for accuracy) . Second, the pixels within the outlined shape of the image are interpolated where the boundary of the image outline expands or contracts. For stereoscopic image conversion, only pixel interpolation horizontal to the image frame axis (indicated by x in FIG. 4) is required, since the second perspective image is a slight horizontal shift from the first perspective image. This simplifies the procedure of image warping and helps speed up the transformation calculations. Indeed, a simple horizontal linear pixel interpolation can be achieved in close to real-time. Objects are the components of an image which are manipulated to give the image depth. Examples of objects are tables, chairs, walls, floors, lights, people, etc. Objects can further be divided into sub-objects to add more depth. For example, a person could be broken into two arm objects, a head object, a body object, and two leg objects.
A frame is a single image at a specific time. NTSC video consists of 30 frames per second or 30 images per second. Sequential frames are the sequential images to be converted.
A key frame is a specific frame or image where input from the user is required to help direct the animation. The start and end of a new scene are key frames, ._.Frames where objects start or stop motion are typically key frames.
Key objects are the important objects in the image which will be processed using this invention.
Clipping occurs when one object (the foreground object) obscures another object (the background object) . The portions of the background object which are obscured are removed or not displayed and the foreground object is displayed instead.
To convert a series of moving images, a "tweening" process preferably is implemented to automate the outlining process. The tweening process involves interpolations of the vector outlined shape of all the outlined objects within the scene from one key frame ("the start key frame") to the next key frame ("the stop key frame"). The interpolation can be either linear or nonlinear (e.g., spline) depending on accuracy and/or speed requirements. Only the end points (i.e., the key frames) of the motion path for the objects need to be defined, significantly reducing the burden on the operator.
In order to apply image warping to the stereoscopic conversion of moving images such as those recorded on motion picture film or video media, two sets of outlined objects are used to enable clipping. A first set of outlines is provided either by an automatic image recognition process and/or manually by an operator to outline vectors on all the necessary key objects within an image. This set of outlines is then shifted and warped through the perspective transformation based on the depth arrangement of the objects provided by the operator. When objects are outlined, a set of vectors are used to represent this. When vectors (outlines) from two objects intersect, vector-crossovers are created. The background vectors should be clipped to conform to the foreground vectors when this happens. Vector cross-overs can result in undesirable effects, therefore, a second set of outlined objects is generated after the perspective transformation.
As shown in FIG. 3, if the coordinates of an object in the left perspective image are crossed with corresponding coordinates in the right perspective image (i.e., each left coordinate is moved to the right of the corresponding right coordinate) , the result is negative parallax and the object appears to protrude toward the observer and away from the neutral plane. Conversely, if the coordinates are uncrossed (each right coordinate is moved to the left of the corresponding left coordinate) , the result is positive parallax and the object appears to recede .
Whenever the foreground objects' vectors cross- over the background objects due to the transformation, the background object's vectors preferably are clipped to the edge of the foreground object's outline. This prevents any vector cross-over during the actual warping process which could cause unwanted warping artifacts. The entire process can be completely automated using conventional computer line clipping algorithms. Given two sets of vectors with one tagged as .the foreground and one tagged as the background, it is well known by someone skilled in the art how to automatically clip the background vectors to the foreground. Algorithms like this are used for hidden line removal in computer graphics. FIG. 1 shows two-dimensional video media images comprism-g unprocessed still frames 2 provided to workstation 3 through digital image capture using video media frame grabber 1. Processed still frames 4 are displayed in sequence on monitor 5.
Key objects are operator-selected foreground objects in a particular sequence of images. Key frames, which correspond to frames m which key objects change motion or deviate from a predicted motion path, can be identified automatically by delta pixel changes between frames. Delta pixel changes can also be used to determine the basic motion path within each scene. The key frames can be automatically selected based on the extreme points of the motion path. The key frames can also be selected by a human operator after viewing the scene. These key frames are then individually processed by outlining the key objects within the image.
Objects can be detected automatically by an edge detection algorithm or manually by an operator. Automatic edge detection algorithms such as Laplacian, Sobel and Prewitt convoluted filters can be used to highlight the object's outline. Such algorithms are described, for example, in Christopher Watkms, Alberto Sadun, Stephen Marenka, "Modern Image Processing: Warping, Morphing, and Classical Techniques," Academic Press Professional, Inc. (1993) . The computer can then trace around the highlighted objects to form vector outlines. A vector is a line which connects points (i.e., the control points) on the object. Once vector outlines are determined for objects within the image, the foreground priority of objects can be determined by first identifying objects which are enclosed within another object or objects which overlap another ^object. These objects are considered to be potential foreground objects and subject to further criteria such as intensity cue and vertical position cue. Key objects can be identified by the operator as corresponding to certain predefined objects for predefined profile warping. Each predefined object preferably has a profile identification which defines its perceived surface profile. This predefined profile warping eliminates the need to define vectors within an object in order to recreate its full volume appearance.
The object's outline defines the area which will be transformed via perspective warping. Perspective warping is a type of image warping in which the image is horizontally shifted. Therefore, the actual transformation is significantly faster to compute because there is no vertical warping.
The object's outline is warped and the object's image within the outlined area is then warped into the necessary perspective. The control points of the outline are transformed into their new shape and position based on the depth arrangements and/or the object's profile identification. The actual transformation only shifts the end points of the vectors horizontally. The amount of the shift may be based, for example, on the computed depth as shown m FIG. 2, which shows the relationship between the distance from the eyes, y, versus the amount of relative shift, Δ, at any horizontal point x for a fixed focus distance of 45 mcnes and an eye separation distance of 3 inches. One should note that this position offset distance saturates rapidly as the depth becomes farther away. A more generalized equation was derived to calculate Δ, the amount of snift required., for the right (i.e., second) perspective image from the left (i.e., first) perspective image at any particular depth relative to the width of the viewing screen, with reference to FIG. 3. The actual derivation was done using a symbolic mathematical program (Maple V) , Maple is a computer algebra system, i.e., a computer program which manipulates mathematical expressions symbolically. Maple is able to solve equations, systems of equations and differential equations, perform differentiation and integration, do high precision arithmetic and draw accurate graphs m two and three dimensions. A listing of the Maple V code is listed m the Appendix which follows this specification and the resulting equation is shown below:
Figure imgf000013_0001
The calculation for the linear pixel interpolation is shown below:
Red.eM[Xne = (l-frac(X) ) •Redold[mteger(X) ] + frac(X) •Redold[ integer (X)+l] Greenneu[Xne = ( 1-frac (X) ) -GreenD .[integer (X) ] + frac(X) 'Greened [integer (X) +1] Blueneu[XneM] = ( 1-frac (X) ) «Blueold [integer (X) ] + frac(X) •Blueold[ nteger(X)+l]
where: Old pixel position.
Old left pixel boundary position.
St Old right pixel boundary position. ne New sampled pixel position. To New left pixel boundary position.
T> New right pixel boundary position. As shown in FIG. 5, the pixels within the boundary- shown m FIG. 5a are resampled to the new boundary shown in FIG. 5b. Since the number of pixels along the x axis is arbitrary, the new sampled pixels are not likely to map back to the original pixels.
For illustration purposes, consider that the left image is the original image and the right image is generated through perspective transformation. The operator will then select a neutral plane within the image. The neutral plane (also referred to as the "zero- shifted" plane) is that plane on which no point shifts between the left and right perspective image. Any point on a plane m front of the neutral plane shifts to the left and any point on a plane behind the neutral plane shifts to the right. The selection of the neutral plane is important to minimize the maximum shift required by the object. The neutral plane can be selected automatically by choosing the center plane between the front-most object and the back-most object. Manual adjustment by the operator is allowed for aesthetic reasons if the operator so chooses.
Objects in front of the neutral plane appear to float out of the viewing screen. This can produce undesirable effects due to object's partial appearance at the border of the screen. This creates an unnatural appearance as the object gets partially clipped by the border behind it instead of m front. This can be corrected by shifting the left and right frames horizontally so that the object is behind the viewing screen.
Other perspective calculation curves, such as the saturating perspective calculation curve shown in FIG. 6, can be used instead of the actual perspective calculation curve shown in FIG. 2. By using a modified perspective calculation curve to perform the perspective transformation, it is possible to avoid excessive image warping and distortion of the transformed image. A similar result can be achieved using the actual perspective calculation curve of FIG. 2 if the operator is prevented from setting the control points too far in front of the neutral plane. Once the main outlines of the object are determined within the scene, the object can be easily transformed into the second perspective image using a perspective transformation curve. To illustrate the technique, the transformation of a simple cube is shown in FIGS. 7a and 7b. FIG. 7a shows the left perspective image. FIG. 7b shows the right perspective image which results from the transformation. In this case, the main outlines are the edges of the cube and the control points are the eight corners of the cube. The control points are manipulated and depth arranged by an operator to determine the current position in depth. This perceived depth is then horizontally translated right to create the second perspective image. The resulting stereoscopic image can be displayed in quasi-real-time on a monitor to provide feedback to the operator.
In this example, image warping based on the vector outline transformation is performed based on the left perspective image. A new set of vectors is generated for use in the warping procedure based on the object's depth arrangement. Any vector outlined object in the background which crosses vectors with the foreground objects will be clipped around the contour of the foreground object to prevent vector crossover during the warping procedure, as shown in FIG. 8. Certain vectors in the new set of vectors are allowed to crossover when the distance of the crossover is sufficiently small, thereby allowing the foreground image to fold over the background image .
Since most video media or motion picture film frames have only minute changes from one frame to the next, it is therefore possible to interpolate (i.e., "tween") the m-between object outlines for frames between the key objects' motion frames.
When motion picture film is converted to video media, the frame rate must be converted from 24 frames per second to 30 frames per second. One common way to do this is by making every fifth video media frame the duplicate of the preceding motion picture film frame. As a way of reducing the number of frames to be processed and increasing the interpolation accuracy, such redundant frames may be removed from video media prior to monoscopic to stereoscopic conversion. Of course, the redundancy must be remtroduced to the video media after processing is complete.
Since there is latency m human visual depth perception, not every video media or motion picture frame m a sequence need be converted to a stereoscopic image. G.R. Engel, "An Investigation of Visual Responses to Brief Stereoscopic Stimuli," 22 Quarterly Journal Of Experimental Psychology, pp. 148-166 (1970). Stereoscopic images only need to be displayed about every 300 milliseconds in order for the entire movie sequence to be a stereoscopic presentation which appears to be continuously stereoscopic. This reduces the computation needed for the conversion of monoscopic motion picture film to ..stereoscopic motion picture film by several orders of magnitude and results in a much quicker and easier conversion. A preferred embodiment of the method of the present invention is shown in FIGS. 9a and 9b. Digital images are captured at step 10. If the digital images are from video media and were originally recorded on motion picture film, redundant frames are removed at step 30. Key objects are selected at step 40. Key frames are identified at step 50. Each frame between the start key frame and the stop key frame is tweened at step 60. Finally, the sequence of stereoscopic images is displayed at step 70. The tweening process of step 60 begins with outlining the objects to form vector outlines at step 100. The depth of objects is set at step 110. If a vector outlined object in the background crosses vectors with a foreground object, the vectors of the background object are clipped at step 130. If the object corresponds to a predefined object it is transformed using a predefined profile warping at step 150. Otherwise, the object is transformed using a selected perspective calculation curve at step 160. Unless a stop key frame is reached, process returns to step 100. The tweening process is repeated for each set of start and stop key frames.
Thus it can be seen that a method and apparatus for converting monoscopic images into stereoscopic images is provided. Although this specification describes the generation of a right perspective image based on a left perspective image, one skilled in the art will appreciate that a left perspective image can just as easily be generated from a right perspective image. Indeed, one skilled in the art will appreciate that the present invention can be practiced by other than the described embodiments, which are presented for purposes of illustration and not of limitation, and the present invention is limited only by the claims which follow.

Claims

WHAT IS CLAIMED IS:
1. A method for converting a monoscopic image comprising a first perspective to a stereoscopic image comprising said first perspective and a second perspective, comprising the steps of: identifying at least one object in said monoscopic image; outlining said at least one object to generate an outline; endowing said at least one object with depth information; and generating said second perspective by image warping sa d at least one object.
2. The method of claim 1 wherein said generating step comprises: shifting points on said outline of said at least one object from said first perspective to said second perspective; and shifting points within said outlιp tron said first perspective to said second persi ective,
3. The method of claim 1 wherein said generating step is performed on a sequence of successive frames of said monoscopic images to produce a stereoscopic presentation.
4. The method of claim 3 wherein said sequence of successive frames is derived from motion picture film.
5. The method of claim 3 wherein said sequence-_-of successive frames is derived from video media.
6. The method of claim 5 wherein said sequence of successive frames contains redundant frames as a result of conversion from motion picture film.
7. The method of claim 6 further comprising a preliminary step of: removing said redundant frames from said sequence of successive frames.
8. The method of claim 3 wherein said method is performed only on selected frames of said sequence of successive frames, thereby reducing the number of frames that must be converted to stereoscopic images, thereby reducing computation time while providing a sequence of images which appear to be stereoscopic.
9. A method for converting a sequence of frames of monoscopic images with at least one foreground object with a depth and an outline and at least one background object with an outline to a sequence of stereoscopic images, comprising the steps of: identifying said at least one foreground object; outlining said at least one foreground object; setting said depth of said at least one foreground object; line clipping said at least one background objeot if said at least one foreground object moves in front of said at least one background object; shifting points of said outline of said at least one foreground object and said outline of said at least one background object to a second image perspective; and shifting points inside said outline of said at least one foreground object and said outline of said at least one background object to a second image perspective.
10. The method of claim 9 wherein said depth in said setting step may be varied in response to instructions from an operator.
11. The method of claim 9 further comprising the step of displaying said sequence of stereoscopic images in near real-time.
12. The method of claim 9 wherein said sequence of successive frames is derived from motion picture film.
13. The method of claim 9 wherein said sequence of successive frames is derived from video media.
14. The method of claim 13 wherein said sequence of successive frames contains redundant frames as a result of conversion from motion picture film.
15. The method of claim 14 further comprising a preliminary step of: removing said redundant frames from said sequence of successive frames.
16. The method of claim 9 wherein said method is performed only on selected frames of said sequence of successive frames, thereby reducing the number of frames that must be converted to stereoscopic images, thereby reducing computation time while providing a sequence of images which appear to be stereoscopic.
17. The method of claim 16 wherein said selected frames are separated by a set interval of time.
18. The method of claim 17 wherein said set interval is about 300 milliseconds.
19. The method of claim 9 further comprising a step before said identifying step of: determining a start key frame and a stop key frame.
20. The method of claim 19 wherein said shifting points of said outline step and said shifting points inside said outline step are repeated for each frame between said start key frame and said stop key frame.
21. The method of claim 19, wherein said shifting points of said outline step and said shifting points inside said outline step are done via non-linear interpolation.
22. The method of claim 20 wherein said shifting points of said outline step and said shifting points inside said outline step are done via linear interpolation.
23. The method of claim 9 wherein said shifting points of said outline step and said shifting points inside said outline step are done based on said depth of said at least one foreground object and a selected perspective calculation curve.
24. The method of claim 23 wherein said selected perspective calculation curve is an actual perspective calculation curve.
25. The method of claim 23 wherein an operator is prevented from setting said depth of said at least one foreground object significantly beyond a neutral plane.
26. The method of claim 23 wherein said selected perspective calculation curve is a modified perspective calculation curve.
27. The method of claim 26 wherein said modified perspective calculation curve is a saturating perspective calculation curve.
28. A method for converting a monoscopic image comprising a first perspective to a stereoscopic image comprising said first perspective and a second perspective, comprising the steps of: identifying at least one object in said monoscopic image; outlining said at least one object; endowing said at least one object with depth information; and generating said second perspective by a method selected from a horizontal shift method, a frame delay method and an image warping method, based on direction of motion of camera and said at least one object, desired computation complexity and desired image quality.
29. A recording medium on which a stereoscopic image comprising a first perspective and a second perspective is recorded, said stereoscopic image being formed from a monoscopic image comprising said first perspective by a method comprising the steps of: identifying at least one object in said monoscopic image; outlining said at least one object; endowing said at least one object with depth information; and generating said second perspective by image warping said at least one object.
30. A product produced by placing on a recording medium a stereoscopic image generated by the method of claims 1, 9 or 28.
31. An apparatus for converting a sequence of frames- of- monoscopic images with at least one foreground object with a depth and at least one background object to a sequence of stereoscopic images, comprising: a foreground object identifier; a foreground object outliner; a foreground object depth setter; a background object line clipper; and a foreground object point shifter coupled in series.
32. The apparatus of claim 31 wherein said sequence of successive frames is derived from motion picture film.
33. The apparatus of claim 31 wherein said sequence of successive frames is derived from video media.
34. The apparatus of claim 33 wherein said sequence of successive frames contains redundant frames as a result of conversion from motion picture film.
35. The apparatus of claim 34 further comprising a redundant frame remover coupled in series before said foreground object identifier.
36. The apparatus of claim 31 wherein said apparatus operates only on selected frames of said monoscopic images, thereby reducing the number of frames that must be converted to stereoscopic images, thereby reducing computation time while providing a sequence of images which appear to be stereoscopic. _ -^37. The apparatus of claim 36 wherein said selected frames are separated by a set interval of time.
38. The apparatus of claim 37 wherein said set interval is about 300 milliseconds.
39. The apparatus of claim 31 wherein said foreground object point shifter shifts points based on said depth of said at least one foreground object and a selected perspective calculation curve.
40. The apparatus of claim 39 wherein said selected perspective calculation curve is an actual perspective calculation curve.
41. The apparatus of claim 39 wherein an operator is prevented' from setting said depth of said at least one foreground object significantly beyond a neutral plane.
42. The apparatus of claim 39 wherein said selected perspective calculation curve is a modified perspective calculation curve.
43. The apparatus of claim 42 wherein said modified perspective calculation curve is a saturating perspective calculation curve.
44. The apparatus of claim 31 wherein said depth in said setting step may be varied in response to instructions from an operator.
.45. The apparatus of claim 44 wherein said sequence of stereoscopic images is displayed in near real- time.
PCT/US1998/012119 1997-07-07 1998-06-10 Method and apparatus for monoscopic to stereoscopic image conversion WO1999003068A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU80668/98A AU8066898A (en) 1997-07-07 1998-06-10 Method and apparatus for monoscopic to stereoscopic image conversion

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/889,713 US6031564A (en) 1997-07-07 1997-07-07 Method and apparatus for monoscopic to stereoscopic image conversion
US08/889,713 1997-07-07

Publications (1)

Publication Number Publication Date
WO1999003068A1 true WO1999003068A1 (en) 1999-01-21

Family

ID=25395653

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1998/012119 WO1999003068A1 (en) 1997-07-07 1998-06-10 Method and apparatus for monoscopic to stereoscopic image conversion

Country Status (3)

Country Link
US (2) US6031564A (en)
AU (1) AU8066898A (en)
WO (1) WO1999003068A1 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2001076258A2 (en) * 2000-04-01 2001-10-11 Tdv Technologies Corp. Generation of a sequence of stereoscopic images from a sequence of 2d images
WO2002013143A1 (en) * 2000-08-04 2002-02-14 Dynamic Digital Depth Research Pty Ltd. Image conversion and encoding technique
US6477267B1 (en) 1995-12-22 2002-11-05 Dynamic Digital Depth Research Pty Ltd. Image conversion and encoding techniques
JP2007502454A (en) * 2003-08-05 2007-02-08 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Multi-view image generation
US7254265B2 (en) 2000-04-01 2007-08-07 Newsight Corporation Methods and systems for 2D/3D image conversion and optimization
WO2007111570A2 (en) * 2006-03-29 2007-10-04 Bracco Imaging S.P.A. Methods and apparatuses for stereoscopic image guided surgical navigation
US7321374B2 (en) 2001-11-24 2008-01-22 Newsight Corporation Method and device for the generation of 3-D images
US7419617B2 (en) 2001-06-13 2008-09-02 Basf Aktiengesellschaft Coolant comprising azole derivatives for cooling systems in fuel-cell drives
CN101783027A (en) * 2010-02-26 2010-07-21 浙江大学 Dynamic scene three-dimensional recording method based on multiple image sensors
GB2475730A (en) * 2009-11-27 2011-06-01 Sony Corp Transformation of occluding objects in 2D to 3D image generation
WO2013109252A1 (en) * 2012-01-17 2013-07-25 Thomson Licensing Generating an image for another view

Families Citing this family (114)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7907793B1 (en) 2001-05-04 2011-03-15 Legend Films Inc. Image sequence depth enhancement system and method
US8396328B2 (en) 2001-05-04 2013-03-12 Legend3D, Inc. Minimal artifact image sequence depth enhancement system and method
US6031564A (en) * 1997-07-07 2000-02-29 Reveo, Inc. Method and apparatus for monoscopic to stereoscopic image conversion
CA2244635C (en) * 1997-08-08 2002-12-10 Kenneth J. Tetterington Method and apparatus to alternate stereoscopic images in a video display device
JP3376868B2 (en) * 1997-08-11 2003-02-10 日産自動車株式会社 Three-dimensional terrain display method for navigation system, navigation system, and medium recording three-dimensional terrain display program
EP2252071A3 (en) * 1997-12-05 2017-04-12 Dynamic Digital Depth Research Pty. Ltd. Improved image conversion and encoding techniques
US6515659B1 (en) 1998-05-27 2003-02-04 In-Three, Inc. Method and system for creating realistic smooth three-dimensional depth contours from two-dimensional images
US6208348B1 (en) * 1998-05-27 2001-03-27 In-Three, Inc. System and method for dimensionalization processing of images in consideration of a pedetermined image projection format
US7116323B2 (en) * 1998-05-27 2006-10-03 In-Three, Inc. Method of hidden surface reconstruction for creating accurate three-dimensional images converted from two-dimensional images
US7116324B2 (en) * 1998-05-27 2006-10-03 In-Three, Inc. Method for minimizing visual artifacts converting two-dimensional motion pictures into three-dimensional motion pictures
US7102633B2 (en) * 1998-05-27 2006-09-05 In-Three, Inc. Method for conforming objects to a common depth perspective for converting two-dimensional images into three-dimensional images
US20050231505A1 (en) * 1998-05-27 2005-10-20 Kaye Michael C Method for creating artifact free three-dimensional images converted from two-dimensional images
US20050146521A1 (en) * 1998-05-27 2005-07-07 Kaye Michael C. Method for creating and presenting an accurate reproduction of three-dimensional images converted from two-dimensional images
US6515662B1 (en) * 1998-07-16 2003-02-04 Canon Kabushiki Kaisha Computer apparatus for providing stereoscopic views from monographic images and method
JP3532772B2 (en) * 1998-09-25 2004-05-31 本田技研工業株式会社 Occupant state detection device
US6388666B1 (en) * 1998-10-27 2002-05-14 Imax Corporation System and method for generating stereoscopic image data
US6295067B1 (en) * 1999-03-12 2001-09-25 Richard Dubnow Method of manufacturing a 3D viewing disk
US7043058B2 (en) * 2001-04-20 2006-05-09 Avid Technology, Inc. Correcting motion vector maps for image processing
US8401336B2 (en) 2001-05-04 2013-03-19 Legend3D, Inc. System and method for rapid image sequence depth enhancement with augmented computer-generated elements
US8897596B1 (en) 2001-05-04 2014-11-25 Legend3D, Inc. System and method for rapid image sequence depth enhancement with translucent elements
US9286941B2 (en) 2001-05-04 2016-03-15 Legend3D, Inc. Image sequence enhancement and motion picture project management system
US7084838B2 (en) * 2001-08-17 2006-08-01 Geo-Rae, Co., Ltd. Method and system for controlling the motion of stereoscopic cameras using a three-dimensional mouse
CA2478671C (en) * 2002-03-13 2011-09-13 Imax Corporation Systems and methods for digitally re-mastering or otherwise modifying motion pictures or other image sequences data
JP3782368B2 (en) * 2002-03-29 2006-06-07 株式会社東芝 Object image clipping method and program, and object image clipping device
US7489812B2 (en) * 2002-06-07 2009-02-10 Dynamic Digital Depth Research Pty Ltd. Conversion and encoding techniques
IL150131A (en) * 2002-06-10 2007-03-08 Rafael Advanced Defense Sys Method for converting a sequence of monoscopic images to a sequence of stereoscopic images
US20040212612A1 (en) * 2003-04-28 2004-10-28 Michael Epstein Method and apparatus for converting two-dimensional images into three-dimensional images
CN100414566C (en) * 2003-06-19 2008-08-27 邓兴峰 Panoramic reconstruction method of three dimensional image from two dimensional image
GB0322840D0 (en) * 2003-09-30 2003-10-29 Lange Eric B Stereoscopic imaging
US7643025B2 (en) * 2003-09-30 2010-01-05 Eric Belk Lange Method and apparatus for applying stereoscopic imagery to three-dimensionally defined substrates
EP1727093A1 (en) * 2003-12-19 2006-11-29 Tdvision Corporation S.A. DE C.V. 3d videogame system
US7342586B2 (en) * 2004-09-13 2008-03-11 Nbor Corporation System and method for creating and playing a tweening animation using a graphic directional indicator
US7542034B2 (en) 2004-09-23 2009-06-02 Conversion Works, Inc. System and method for processing video images
IL166305A0 (en) * 2005-01-14 2006-01-15 Rafael Armament Dev Authority Automatic conversion from monoscopic video to stereoscopic video
WO2006078237A1 (en) * 2005-01-15 2006-07-27 In-Three, Inc. Method for converting two-dimensional images to three-dimensional images using depth perspective
GB0502891D0 (en) 2005-02-12 2005-03-16 Next Device Ltd User interfaces
US8004515B1 (en) * 2005-03-15 2011-08-23 Nvidia Corporation Stereoscopic vertex shader override
EP1889225A4 (en) 2005-06-03 2012-05-16 Mediapod Multi-dimensional imaging system and method
US20070127909A1 (en) 2005-08-25 2007-06-07 Craig Mowry System and apparatus for increasing quality and efficiency of film capture and methods of use thereof
EP1938136A2 (en) * 2005-10-16 2008-07-02 Mediapod LLC Apparatus, system and method for increasing quality of digital image capture
JP4887727B2 (en) * 2005-10-20 2012-02-29 ソニー株式会社 Image signal processing apparatus, camera system, and image signal processing method
JP5059024B2 (en) * 2005-12-19 2012-10-24 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ 3D image display method and apparatus
CN101375315B (en) * 2006-01-27 2015-03-18 图象公司 Methods and systems for digitally re-mastering of 2D and 3D motion pictures for exhibition with enhanced visual quality
AU2006201851A1 (en) * 2006-05-03 2007-11-22 Canon Kabushiki Kaisha Compressing page descriptions while preserving high quality
WO2007148219A2 (en) 2006-06-23 2007-12-27 Imax Corporation Methods and systems for converting 2d motion pictures for stereoscopic 3d exhibition
US8655052B2 (en) * 2007-01-26 2014-02-18 Intellectual Discovery Co., Ltd. Methodology for 3D scene reconstruction from 2D image sequences
US20080225040A1 (en) * 2007-03-12 2008-09-18 Conversion Works, Inc. System and method of treating semi-transparent features in the conversion of two-dimensional images to three-dimensional images
US20080225042A1 (en) * 2007-03-12 2008-09-18 Conversion Works, Inc. Systems and methods for allowing a user to dynamically manipulate stereoscopic parameters
US8274530B2 (en) 2007-03-12 2012-09-25 Conversion Works, Inc. Systems and methods for filling occluded information for 2-D to 3-D conversion
US8488868B2 (en) * 2007-04-03 2013-07-16 Her Majesty The Queen In Right Of Canada, As Represented By The Minister Of Industry, Through The Communications Research Centre Canada Generation of a depth map from a monoscopic color image for rendering stereoscopic still and video images
US8213711B2 (en) * 2007-04-03 2012-07-03 Her Majesty The Queen In Right Of Canada As Represented By The Minister Of Industry, Through The Communications Research Centre Canada Method and graphical user interface for modifying depth maps
EP2168384A4 (en) * 2007-06-19 2013-11-06 Korea Electronics Telecomm Metadata structure for storing and playing stereoscopic data, and method for storing stereoscopic content file using this metadata
US9196080B2 (en) 2008-02-05 2015-11-24 Disney Enterprises, Inc. Medial axis decomposition of 2D objects to synthesize binocular depth
US8471898B2 (en) * 2008-02-05 2013-06-25 Disney Enterprises, Inc. Medial axis decomposition of 2D objects to synthesize binocular depth
BRPI0914466A2 (en) * 2008-10-28 2015-10-27 Koninkl Philips Electronics Nv method of generating an occlusion image property map for an occlusion viewing position for a three-dimensional scene, computer program product, software tool for use with a three-dimensional modeling computer program to generate a three-dimensional property map occlusion imaging and apparatus for generating an occlusion image property map for an occlusion viewing position for a three-dimensional scene
US8233664B2 (en) * 2008-11-12 2012-07-31 Eastman Kodak Company Determining relative depth of points in multiple videos
US20100220893A1 (en) * 2009-03-02 2010-09-02 Gwo Giun Lee Method and System of Mono-View Depth Estimation
JP5627860B2 (en) * 2009-04-27 2014-11-19 三菱電機株式会社 3D image distribution system, 3D image distribution method, 3D image distribution device, 3D image viewing system, 3D image viewing method, 3D image viewing device
US9380292B2 (en) 2009-07-31 2016-06-28 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating three-dimensional (3D) images of a scene
US20110025830A1 (en) * 2009-07-31 2011-02-03 3Dmedia Corporation Methods, systems, and computer-readable storage media for generating stereoscopic content via depth map creation
WO2011014419A1 (en) * 2009-07-31 2011-02-03 3Dmedia Corporation Methods, systems, and computer-readable storage media for creating three-dimensional (3d) images of a scene
US8254760B2 (en) * 2009-08-28 2012-08-28 Apple Inc. Pixel analysis and frame alignment for background frames
US8624959B1 (en) * 2009-09-11 2014-01-07 The Boeing Company Stereo video movies
JP4763822B2 (en) * 2009-09-30 2011-08-31 パナソニック株式会社 Video signal processing apparatus and video signal processing method
CN101697235B (en) * 2009-10-30 2013-04-10 青岛海信移动通信技术股份有限公司 Perspective drawing generation method and perspective drawing generation device
JP5478268B2 (en) * 2010-01-13 2014-04-23 任天堂株式会社 Image processing program, image processing apparatus, image processing method, and image processing system
US8963996B2 (en) * 2010-05-05 2015-02-24 Samsung Electronics Co., Ltd. Communication of stereoscopic three-dimensional (3D) video information including an uncompressed eye view video frames
US8421847B2 (en) * 2010-05-21 2013-04-16 Mediatek Inc. Apparatus and method for converting two-dimensional video frames to stereoscopic video frames
US9132352B1 (en) 2010-06-24 2015-09-15 Gregory S. Rabin Interactive system and method for rendering an object
US9053562B1 (en) * 2010-06-24 2015-06-09 Gregory S. Rabin Two dimensional to three dimensional moving image converter
US9344701B2 (en) 2010-07-23 2016-05-17 3Dmedia Corporation Methods, systems, and computer-readable storage media for identifying a rough depth map in a scene and for determining a stereo-base distance for three-dimensional (3D) content creation
US8831273B2 (en) 2010-09-10 2014-09-09 Reald Inc. Methods and systems for pre-processing two-dimensional image files to be converted to three-dimensional image files
US9485497B2 (en) 2010-09-10 2016-11-01 Reald Inc. Systems and methods for converting two-dimensional images into three-dimensional images
US9185388B2 (en) 2010-11-03 2015-11-10 3Dmedia Corporation Methods, systems, and computer program products for creating three-dimensional video sequences
US8274552B2 (en) 2010-12-27 2012-09-25 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
US10200671B2 (en) 2010-12-27 2019-02-05 3Dmedia Corporation Primary and auxiliary image capture devices for image processing and related methods
JP5502211B2 (en) * 2011-01-17 2014-05-28 パナソニック株式会社 Stereoscopic image processing apparatus and stereoscopic image processing method
US8730232B2 (en) 2011-02-01 2014-05-20 Legend3D, Inc. Director-style based 2D to 3D movie conversion system and method
US9288476B2 (en) 2011-02-17 2016-03-15 Legend3D, Inc. System and method for real-time depth modification of stereo images of a virtual reality environment
US9241147B2 (en) 2013-05-01 2016-01-19 Legend3D, Inc. External depth map transformation method for conversion of two-dimensional images to stereoscopic images
US9282321B2 (en) 2011-02-17 2016-03-08 Legend3D, Inc. 3D model multi-reviewer system
US9113130B2 (en) 2012-02-06 2015-08-18 Legend3D, Inc. Multi-stage production pipeline system
US9407904B2 (en) 2013-05-01 2016-08-02 Legend3D, Inc. Method for creating 3D virtual reality from 2D images
TR201103444A2 (en) 2011-04-08 2012-10-22 Vestel Elektron�K Sanay� Ve T�Caret A.�. Method and device for creating a 3D image from a 2D image.
CN102903143A (en) * 2011-07-27 2013-01-30 国际商业机器公司 Method and system for converting two-dimensional image into three-dimensional image
US20130050414A1 (en) * 2011-08-24 2013-02-28 Ati Technologies Ulc Method and system for navigating and selecting objects within a three-dimensional video image
KR101859412B1 (en) * 2011-09-05 2018-05-18 삼성전자 주식회사 Apparatus and method for converting 2d content into 3d content
EP2812894A4 (en) * 2012-02-06 2016-04-06 Legend3D Inc Motion picture project management system
KR101318552B1 (en) * 2012-03-12 2013-10-16 가톨릭대학교 산학협력단 Method for measuring recognition warping about 3d image
US20130257851A1 (en) * 2012-04-01 2013-10-03 Chao-Hua Lee Pipeline web-based process for 3d animation
US9007365B2 (en) 2012-11-27 2015-04-14 Legend3D, Inc. Line depth augmentation system and method for conversion of 2D images to 3D images
US9547937B2 (en) 2012-11-30 2017-01-17 Legend3D, Inc. Three-dimensional annotation system and method
US9215448B2 (en) 2013-01-31 2015-12-15 Here Global B.V. Stereo panoramic images
US9007404B2 (en) 2013-03-15 2015-04-14 Legend3D, Inc. Tilt-based look around effect image enhancement method
US9438878B2 (en) 2013-05-01 2016-09-06 Legend3D, Inc. Method of converting 2D video to 3D video using 3D object models
US9443281B2 (en) * 2014-06-27 2016-09-13 Intel Corporation Pixel-based warping and scaling accelerator
US10262426B2 (en) 2014-10-31 2019-04-16 Fyusion, Inc. System and method for infinite smoothing of image sequences
US10650574B2 (en) * 2014-10-31 2020-05-12 Fyusion, Inc. Generating stereoscopic pairs of images from a single lens camera
US9940541B2 (en) 2015-07-15 2018-04-10 Fyusion, Inc. Artificially rendering images using interpolation of tracked control points
US10176592B2 (en) 2014-10-31 2019-01-08 Fyusion, Inc. Multi-directional structured image array capture on a 2D graph
US10275935B2 (en) 2014-10-31 2019-04-30 Fyusion, Inc. System and method for infinite synthetic image generation from multi-directional structured image array
EP3086554B1 (en) * 2015-04-24 2019-04-24 Visual Vertigo Software Technologies GmbH System and method for producing and dispensing stereoscopic video films
US11095869B2 (en) 2015-09-22 2021-08-17 Fyusion, Inc. System and method for generating combined embedded multi-view interactive digital media representations
US10852902B2 (en) 2015-07-15 2020-12-01 Fyusion, Inc. Automatic tagging of objects on a multi-view interactive digital media representation of a dynamic entity
US10147211B2 (en) 2015-07-15 2018-12-04 Fyusion, Inc. Artificially rendering images using viewpoint interpolation and extrapolation
US10242474B2 (en) 2015-07-15 2019-03-26 Fyusion, Inc. Artificially rendering images using viewpoint interpolation and extrapolation
US10222932B2 (en) 2015-07-15 2019-03-05 Fyusion, Inc. Virtual reality environment based manipulation of multilayered multi-view interactive digital media representations
US9609307B1 (en) 2015-09-17 2017-03-28 Legend3D, Inc. Method of converting 2D video to 3D video using machine learning
US11783864B2 (en) 2015-09-22 2023-10-10 Fyusion, Inc. Integration of audio into a multi-view interactive digital media representation
US10003732B2 (en) 2016-02-25 2018-06-19 Foodim Ltd Depth of field processing
US11202017B2 (en) 2016-10-06 2021-12-14 Fyusion, Inc. Live style transfer on a mobile device
US10313651B2 (en) 2017-05-22 2019-06-04 Fyusion, Inc. Snapshots at predefined intervals or angles
US11069147B2 (en) 2017-06-26 2021-07-20 Fyusion, Inc. Modification of multi-view interactive digital media representation
US10592747B2 (en) 2018-04-26 2020-03-17 Fyusion, Inc. Method and apparatus for 3-D auto tagging

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4925294A (en) * 1986-12-17 1990-05-15 Geshwind David M Method to convert two dimensional motion pictures for three-dimensional systems
US5510832A (en) * 1993-12-01 1996-04-23 Medi-Vision Technologies, Inc. Synthesized stereoscopic imaging system and method
EP0735512A2 (en) * 1995-03-29 1996-10-02 SANYO ELECTRIC Co., Ltd. Methods for creating an image for a three-dimensional display, for calculating depth information, and for image processing using the depth information
EP0798666A2 (en) * 1991-09-27 1997-10-01 Sony Corporation Image processing apparatus
US5748199A (en) * 1995-12-20 1998-05-05 Synthonics Incorporated Method and apparatus for converting a two dimensional motion picture into a three dimensional motion picture

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4645459A (en) * 1982-07-30 1987-02-24 Honeywell Inc. Computer generated synthesized imagery
US5050984A (en) * 1983-05-09 1991-09-24 Geshwind David M Method for colorizing footage
GB2210540A (en) * 1987-09-30 1989-06-07 Philips Electronic Associated Method of and arrangement for modifying stored data,and method of and arrangement for generating two-dimensional images
GB8827952D0 (en) * 1988-11-30 1989-01-05 Screen Form Inc Display device
US5973700A (en) * 1992-09-16 1999-10-26 Eastman Kodak Company Method and apparatus for optimizing the resolution of images which have an apparent depth
FR2700654B1 (en) * 1993-01-19 1995-02-17 Thomson Csf Method for estimating disparity between monoscopic images constituting a stereoscopic image.
US5479603A (en) * 1993-07-21 1995-12-26 Xerox Corporation Method and apparatus for producing a composite second image in the spatial context of a first image
US5652851A (en) * 1993-07-21 1997-07-29 Xerox Corporation User interface technique for producing a second image in the spatial context of a first image using a model-based operation
US5581638A (en) * 1993-07-26 1996-12-03 E-Systems, Inc. Method for autonomous image registration
US5523886A (en) * 1994-01-04 1996-06-04 Sega Of America, Inc. Stereoscopic/monoscopic video display system
US5511153A (en) * 1994-01-18 1996-04-23 Massachusetts Institute Of Technology Method and apparatus for three-dimensional, textured models from plural video images
US5510831A (en) * 1994-02-10 1996-04-23 Vision Iii Imaging, Inc. Autostereoscopic imaging apparatus and method using suit scanning of parallax images
US6005967A (en) * 1994-02-18 1999-12-21 Matushita Electric Industrial Co., Ltd. Picture synthesizing apparatus and method
JP2846840B2 (en) * 1994-07-14 1999-01-13 三洋電機株式会社 Method for generating 3D image from 2D image
US6031564A (en) * 1997-07-07 2000-02-29 Reveo, Inc. Method and apparatus for monoscopic to stereoscopic image conversion

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4925294A (en) * 1986-12-17 1990-05-15 Geshwind David M Method to convert two dimensional motion pictures for three-dimensional systems
EP0798666A2 (en) * 1991-09-27 1997-10-01 Sony Corporation Image processing apparatus
US5510832A (en) * 1993-12-01 1996-04-23 Medi-Vision Technologies, Inc. Synthesized stereoscopic imaging system and method
EP0735512A2 (en) * 1995-03-29 1996-10-02 SANYO ELECTRIC Co., Ltd. Methods for creating an image for a three-dimensional display, for calculating depth information, and for image processing using the depth information
US5748199A (en) * 1995-12-20 1998-05-05 Synthonics Incorporated Method and apparatus for converting a two dimensional motion picture into a three dimensional motion picture

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6477267B1 (en) 1995-12-22 2002-11-05 Dynamic Digital Depth Research Pty Ltd. Image conversion and encoding techniques
US7999844B2 (en) 1995-12-22 2011-08-16 Dynamic Digital Depth Research Pty Ltd. Image conversion and encoding techniques
WO2001076258A3 (en) * 2000-04-01 2002-09-12 Tdv Technologies Corp Generation of a sequence of stereoscopic images from a sequence of 2d images
WO2001076258A2 (en) * 2000-04-01 2001-10-11 Tdv Technologies Corp. Generation of a sequence of stereoscopic images from a sequence of 2d images
US7254264B2 (en) 2000-04-01 2007-08-07 Newsight Corporation Method and device for generating 3D images
US7254265B2 (en) 2000-04-01 2007-08-07 Newsight Corporation Methods and systems for 2D/3D image conversion and optimization
WO2002013143A1 (en) * 2000-08-04 2002-02-14 Dynamic Digital Depth Research Pty Ltd. Image conversion and encoding technique
US7419617B2 (en) 2001-06-13 2008-09-02 Basf Aktiengesellschaft Coolant comprising azole derivatives for cooling systems in fuel-cell drives
US7419618B2 (en) 2001-06-13 2008-09-02 Basf Aktiengesellschaft Coolant comprising azole derivatives for cooling systems in fuel-cell drives
US7321374B2 (en) 2001-11-24 2008-01-22 Newsight Corporation Method and device for the generation of 3-D images
JP2007502454A (en) * 2003-08-05 2007-02-08 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Multi-view image generation
JP4644669B2 (en) * 2003-08-05 2011-03-02 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Multi-view image generation
WO2007111570A3 (en) * 2006-03-29 2008-06-05 Bracco Imaging Spa Methods and apparatuses for stereoscopic image guided surgical navigation
WO2007111570A2 (en) * 2006-03-29 2007-10-04 Bracco Imaging S.P.A. Methods and apparatuses for stereoscopic image guided surgical navigation
GB2475730A (en) * 2009-11-27 2011-06-01 Sony Corp Transformation of occluding objects in 2D to 3D image generation
US8509521B2 (en) 2009-11-27 2013-08-13 Sony Corporation Method and apparatus and computer program for generating a 3 dimensional image from a 2 dimensional image
CN101783027A (en) * 2010-02-26 2010-07-21 浙江大学 Dynamic scene three-dimensional recording method based on multiple image sensors
WO2013109252A1 (en) * 2012-01-17 2013-07-25 Thomson Licensing Generating an image for another view

Also Published As

Publication number Publication date
US6215516B1 (en) 2001-04-10
US6031564A (en) 2000-02-29
AU8066898A (en) 1999-02-08

Similar Documents

Publication Publication Date Title
US6215516B1 (en) Method and apparatus for monoscopic to stereoscopic image conversion
JP5567578B2 (en) Method and system for processing an input 3D video signal
US6175379B1 (en) Stereoscopic CG image generating apparatus and stereoscopic TV apparatus
EP2340534B1 (en) Optimal depth mapping
US8922628B2 (en) System and process for transforming two-dimensional images into three-dimensional images
US4925294A (en) Method to convert two dimensional motion pictures for three-dimensional systems
US20160182879A1 (en) Critical alignment of parallax images for autostereoscopic display
US9087406B2 (en) Automated stereoscopic computer-animation techniques for determining scaled stereo parameters
WO2010084724A1 (en) Image processing device, program, image processing method, recording method, and recording medium
US20020191841A1 (en) Image processing method and apparatus
US10136121B2 (en) System, method and software for producing virtual three dimensional images that appear to project forward of or above an electronic display
US6252982B1 (en) Image processing system for handling depth information
Berretty et al. Real-time rendering for multiview autostereoscopic displays
JP5396877B2 (en) Image processing apparatus, program, image processing method, and recording method
JPH07200870A (en) Stereoscopic three-dimensional image generator
JP5307051B2 (en) Stereoscopic image adjusting apparatus and adjusting method
KR101121979B1 (en) Method and device for stereoscopic image conversion
JP3091644B2 (en) 3D image conversion method for 2D images
KR20020037097A (en) The apparatus and Method for Vergence Control of a Parallel-Axis Camera using Signal Processing
JP3802629B2 (en) Stereo image generator
Jeong et al. Depth image‐based rendering for multiview generation
GB2289819A (en) Generating autosteroscopic images
JP2975837B2 (en) Method for converting a part of a two-dimensional image to a three-dimensional image
Rhee et al. Stereoscopic view synthesis by view morphing
Kim et al. Resizing of stereoscopic images for display adaptation

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GE GH GM GW HU ID IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG UZ VN YU ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR NE SN TD TG

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: KR

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: CA