WO2002027667A1 - Method for automated two-dimensional and three-dimensional conversion - Google Patents

Method for automated two-dimensional and three-dimensional conversion Download PDF

Info

Publication number
WO2002027667A1
WO2002027667A1 PCT/US2001/028563 US0128563W WO0227667A1 WO 2002027667 A1 WO2002027667 A1 WO 2002027667A1 US 0128563 W US0128563 W US 0128563W WO 0227667 A1 WO0227667 A1 WO 0227667A1
Authority
WO
WIPO (PCT)
Prior art keywords
images
image
dimensional
image file
border
Prior art date
Application number
PCT/US2001/028563
Other languages
French (fr)
Inventor
William M. Karszes
Jerry C. Nims
Paul F. Peters
Tom Bringham
Original Assignee
Orasee Corp.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Orasee Corp. filed Critical Orasee Corp.
Priority to AU2001290838A priority Critical patent/AU2001290838A1/en
Priority to JP2002531371A priority patent/JP2004510272A/en
Priority to EP01970887A priority patent/EP1323135A1/en
Publication of WO2002027667A1 publication Critical patent/WO2002027667A1/en

Links

Classifications

    • G06T5/77
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/261Image signal generators with monoscopic-to-stereoscopic image conversion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/111Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/189Recording image signals; Reproducing recorded image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/286Image signal generators having separate monoscopic and stereoscopic modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays
    • H04N13/305Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays using lenticular lenses, e.g. arrangements of cylindrical lenses
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0077Colour aspects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0092Image segmentation from stereoscopic image signals

Definitions

  • the present invention relates to image processing and fixation and, more particularly, to a method for converting a two-dimensional image to a multiple images and interphasing back to a single image for fixation and display as a three-dimensional image.
  • An example embodiment of the present invention includes a step of receiving a two-dimensional pixel image data file into a data storage, the two-dimensional image file having pixel data reflecting a plurality of objects, and a step retrieving the two-dimensional pixel image data file and a step of displaying an image corresponding to the image file on a video display.
  • the next step receives segmentation command indicating an initial border for at least one of the objects.
  • the next step receives a border refinement command having parameters for generating a final border for at least one of the objects, the parameters being from a group consisting of one or more of color, noise frequency, and edge softness .
  • a following step of the first example embodiment segments the image file into a plurality of images based on the segmentation command and the border refinement command.
  • in-painting of at least one of the plurality of images is performed which fills a portion of the at least one image with a pixels generated based on pixels within the object represented by the at least one image file.
  • a subsequent step receives a first depth data and second depth data for at least a first and a second of the plurality images.
  • a parallax shift step generates at least two phase-shifted images for each of the at least first and second of the plurality of images, the two phase-shifted images having a parallax shift with respect to one another and with respect to at least one other of the plurality of images, the parallax shift being based on the depth data for that image.
  • the plurality of phase-shifted images are interphased into a single interphased image file for output and printing.
  • a further aspect of a method according to this invention includes a further step of receiving a plurality of two-dimensional image files, and overlaying these into a single two-dimensional image file for the segmentation step.
  • the present invention provides a software package controlled by optical principles for the conversion of a single digital input image to multi images and then interphasing back to a single master image.
  • the new image created a program has intensified information.
  • This image when viewed through optical material, has a large variety of visual uses and applications including but not limited to three-dimensional, flip, morph, zoom and action.
  • the same interphasing segment can be used when multiple images are input to produce three dimensions from multiple image cameras in the case of different or sequential images, flip, morph, zoom or action images are produced .
  • FIG. 1 is a perspective view of a first example two- object corresponding to pixel image file
  • FIGS. 1.1, 1.2, 1.3 and 1.4 show an example first, second, third and fourth parallax-shifted image, corresponding to a parallax shifting step according to the present invention of the FIG. 1 image ;
  • FIG. 1.5 shows an interphased image based on interphasing according to the present invention of the FIGS. 1.1 through 1.4 images,- [0012]
  • FIG. 1.6 shows a close up of an example portion of the interphased image shown in FIG. 1.5;
  • FIG. 2.0 shows the FIG. 1.5 example image fixed upon a media
  • FIG. 2.1 shows an example foreground object from within the example image shown in FIG. 1 as fixed within the FIG. 2.0 image ;
  • FIG. 2.2 shows an example background object from within the example image shown in FIG. 1 as fixed within the FIG. 2.0 image ;
  • FIG. 2.3 depicts an example thin film multiple lens sheet for overlay on the FIG. 2.0 fixed image
  • FIG. 2.4 shows an example image as would be seen by a human observer of the FIG. 2.0 image seen through the FIG. 2.3 multiple lens overlay;
  • FIG. 2.5 is a graphical representation of the eyes of the observer in FIG. 2.4;
  • FIGS. 3.1 - 3.5 depict an example segmentation and border refinement step within the method of this invention, for segmenting the two-dimensional image into a plurality of objects;
  • FIGS. 4.1 - 4.4 show a gap in the background object image corresponding to a plurality of parallax shifts in accordance with the method of this invention;
  • FIGS. 5.1 - 5.4 depict an inpainting step performed by the method of this invention for filling the gaps depicted in FIGS. 4.1 - 4.4; of a welded mount double clip according to the present invention.
  • One software package of the present invention effects conversion of a single digital input, two- dimensional image FIG. 1 to multi-images FIG. 1.1-1.4 and then interphases the images 1.1-1.4 back to a single master three-dimensional image FIG. 1.5 having intensified the information.
  • Image 1.5 can be output as a photograph through the use of special photographic printers or printed out directly on standard PC printers and master you can clearly see each of the 4 input images FIG. 1.1-1.4.
  • FIG. 2.0 displaying objects 2.1 and 2.2 through MOM FIG. 2.3
  • the resulting image FIG. 2.4 appears to the observer FIG. 2.5 as having three-dimensions, actions, flip or zoom.
  • FIG. 1.5 the following procedure is preferably followed.
  • the initial two-dimensional image 1 and then subjected to a preparation phase, wherein the image undergoes segmenting involving the masking of each object FIG. 2.1-2.2 in the image by an edge detection method, e.g. "intelligent scissors". Magnetic Lasso of FIG. 3.1-3.4.
  • the resulting object FIG 3.5 is segmented onto a separate layer.
  • Segmenting is the process by which the singe 2D image is separated into layers based on depth in the scene. This allows the layers to be shifted by different amounts to simulate parallax.
  • the users first draws a rough border around the object to be separated FIG. 3.1-3.4.
  • the user can dynamically refine the area without having to paint by hand. These adjustments can be refined for the whole image or just a particular region to allow more accurate segmentation.
  • the parameters are color (hue, saturation, and brightness) , noise frequency and edge softness.
  • the user can adjust the rough point the user can "bake” or apply the parameters to the rough border which results in a new, more accurate rough edge border which can further refined using the adjustable parameters or through hand painting in perspective and moving resulting in images FIG. 1.1-1.4.
  • the images undergo "in-Painting" FIG. 4.1-4.4 (involving the painting and removal of holes that were created during the segmented step) .
  • the algorithms calculates the pixels that will require filling. This is done through intersection and subtraction calculations on the masks or alpha channels of the layers above the current layer.
  • the software then uses a scanline algorithm, only filling one pixel per intersection per scanline until the area is filled.
  • the color of each pixel is based on a weighted average of any surrounding pixels that already have color either from the original image or from earlier passes of this algorithm.
  • the weight of each pixel is adjusted after each pass so that it relies evenly on top/bottom and right/left pixels for the first passes, and relies more heavily on the right/left pixels as the distance to original pixels increases.
  • the user selects the depth by entering the value into the GUI edit field provided.
  • the software assigns layer order by taking the layers in the order in which they were selected.
  • the software then accurately generates the frames necessary for the interphasing step. See FIG. 1.1- 1.4.
  • [0032] Can be used when multiple images are input to produce three-dimensions from multiple image cameras, the case of different or sequential images, flip, morph, zoom or action images are produced.
  • the software package of this invention is controlled by the same optical principles used in three-dimensional photography. Such principles are described, for example in US Patent Nos . 3,852,787; 3,895,867; and 3,960,563; each of the references being hereby incorporated by reference herein in the entirety.

Abstract

The present invention provides a method for controlling known optical principles for the conversion of a single master image (1) into a plurality of images, performing parallax shift (1.1-1.4), and interphasing the images into a single, master image (1.5). The master image when viewed through a MOM (2.3), has a large variety of visual uses and applications including but not limited to three-dimensional, flip, morph, zoom and action.

Description

METHOD FOR AUTOMATED TWO-DIMENSIONAL AND THREE- DIMENSIONAL CONVERSION
BACKGROUND OF THE INVENTION
[0001] Priority of this application is based on U.S. Provisional Application No. 60/232,410, filed on September 14, 2000, which is hereby incorporated by reference.
Field of the Invention
[0002] The present invention relates to image processing and fixation and, more particularly, to a method for converting a two-dimensional image to a multiple images and interphasing back to a single image for fixation and display as a three-dimensional image.
Statement of the Related Art
[0003] Three-dimensional photography is well-known but has presented problems relating to its need for special cameras, special processing, and trained personnel. Multiple lens overlays have been used for display of images having an appearance of three dimensions, but these have the same shortcomings, namely, a requirement for special equipment and specially trained personnel .
[0004] Accordingly, a need has been recognized by the present inventors for a making three-dimensional viewing of existing two-dimensional images more accessible and economical while, at the same time, requiring minimal skill for generating quality images. Preferably, the method would utilize a conventional general purpose programmable computer ("PC"), without any hardware changes .
Summary of the Invention
[0005] An example embodiment of the present invention includes a step of receiving a two-dimensional pixel image data file into a data storage, the two-dimensional image file having pixel data reflecting a plurality of objects, and a step retrieving the two-dimensional pixel image data file and a step of displaying an image corresponding to the image file on a video display. The next step receives segmentation command indicating an initial border for at least one of the objects. The next step receives a border refinement command having parameters for generating a final border for at least one of the objects, the parameters being from a group consisting of one or more of color, noise frequency, and edge softness .
[0006] A following step of the first example embodiment segments the image file into a plurality of images based on the segmentation command and the border refinement command. Next, in-painting of at least one of the plurality of images is performed which fills a portion of the at least one image with a pixels generated based on pixels within the object represented by the at least one image file. A subsequent step receives a first depth data and second depth data for at least a first and a second of the plurality images. Next, a parallax shift step generates at least two phase-shifted images for each of the at least first and second of the plurality of images, the two phase-shifted images having a parallax shift with respect to one another and with respect to at least one other of the plurality of images, the parallax shift being based on the depth data for that image. Next, the plurality of phase-shifted images are interphased into a single interphased image file for output and printing.
[0007] A further aspect of a method according to this invention includes a further step of receiving a plurality of two-dimensional image files, and overlaying these into a single two-dimensional image file for the segmentation step.
SUMMARY OF THE INVENTION
Application of the principles of optics and three- dimensional photograph in a computer program algorithm shows the needs described above to be fulfilled. The ramifications of this program extend to many disciples such as medicines, advertising, photography, novelties, virtual reality hard copy, and the like. The present invention provides a software package controlled by optical principles for the conversion of a single digital input image to multi images and then interphasing back to a single master image. The new image created a program has intensified information. This image when viewed through optical material, has a large variety of visual uses and applications including but not limited to three-dimensional, flip, morph, zoom and action.
[0008] In addition, with the software program of this invention, the same interphasing segment can be used when multiple images are input to produce three dimensions from multiple image cameras in the case of different or sequential images, flip, morph, zoom or action images are produced .
Brief Description of the Drawings
[0009] FIG. 1 is a perspective view of a first example two- object corresponding to pixel image file;
[0010] FIGS. 1.1, 1.2, 1.3 and 1.4 show an example first, second, third and fourth parallax-shifted image, corresponding to a parallax shifting step according to the present invention of the FIG. 1 image ;
[0011] FIG. 1.5 shows an interphased image based on interphasing according to the present invention of the FIGS. 1.1 through 1.4 images,- [0012] FIG. 1.6 shows a close up of an example portion of the interphased image shown in FIG. 1.5;
[0013] FIG. 2.0 shows the FIG. 1.5 example image fixed upon a media;
[0014] FIG. 2.1 shows an example foreground object from within the example image shown in FIG. 1 as fixed within the FIG. 2.0 image ;
[0015] FIG. 2.2 shows an example background object from within the example image shown in FIG. 1 as fixed within the FIG. 2.0 image ;
[0016] FIG. 2.3 depicts an example thin film multiple lens sheet for overlay on the FIG. 2.0 fixed image;
[0017] FIG. 2.4 shows an example image as would be seen by a human observer of the FIG. 2.0 image seen through the FIG. 2.3 multiple lens overlay;
[0018] FIG. 2.5 is a graphical representation of the eyes of the observer in FIG. 2.4;
[0019] FIGS. 3.1 - 3.5 depict an example segmentation and border refinement step within the method of this invention, for segmenting the two-dimensional image into a plurality of objects; [0020] FIGS. 4.1 - 4.4 show a gap in the background object image corresponding to a plurality of parallax shifts in accordance with the method of this invention;
[0021] FIGS. 5.1 - 5.4 depict an inpainting step performed by the method of this invention for filling the gaps depicted in FIGS. 4.1 - 4.4; of a welded mount double clip according to the present invention.
Detailed Description of the Invention
[0022] With the advent of personal computers and high storage capabilities, it has become possible to convert single images into multiple images and then interphase these images back into a master file. This file can be output as a photograph through the use of special photographic principles combined with MOM to produce multi-image outputs, and/or multi-layered outputs with intensified visual data. As illustrated in FIGS. 1.1-4.4.
[0023] One software package of the present invention (not shown) effects conversion of a single digital input, two- dimensional image FIG. 1 to multi-images FIG. 1.1-1.4 and then interphases the images 1.1-1.4 back to a single master three-dimensional image FIG. 1.5 having intensified the information. Image 1.5 can be output as a photograph through the use of special photographic printers or printed out directly on standard PC printers and master you can clearly see each of the 4 input images FIG. 1.1-1.4. [0024] As shown below when FIG. 2.0 displaying objects 2.1 and 2.2 through MOM FIG. 2.3 the resulting image FIG. 2.4 appears to the observer FIG. 2.5 as having three-dimensions, actions, flip or zoom.
[0025] FIG. 1.5 the following procedure is preferably followed. The initial two-dimensional image 1 and then subjected to a preparation phase, wherein the image undergoes segmenting involving the masking of each object FIG. 2.1-2.2 in the image by an edge detection method, e.g. "intelligent scissors". Magnetic Lasso of FIG. 3.1-3.4. The resulting object FIG 3.5 is segmented onto a separate layer.
[0026] Segmenting is the process by which the singe 2D image is separated into layers based on depth in the scene. This allows the layers to be shifted by different amounts to simulate parallax.
[0027] To begin the process of segmenting, the users first draws a rough border around the object to be separated FIG. 3.1-3.4. Next, by adjusting the threshold parameters to the segmenting algorithm, the user can dynamically refine the area without having to paint by hand. These adjustments can be refined for the whole image or just a particular region to allow more accurate segmentation. The parameters are color (hue, saturation, and brightness) , noise frequency and edge softness. The user can adjust the rough point the user can "bake" or apply the parameters to the rough border which results in a new, more accurate rough edge border which can further refined using the adjustable parameters or through hand painting in perspective and moving resulting in images FIG. 1.1-1.4. Next the images undergo "in-Painting" FIG. 4.1-4.4 (involving the painting and removal of holes that were created during the segmented step) .
[0028] When the layers are shifted to simulate parallax, blank areas on the layers below will become possible. Our software will fill in these blanks "In-Painting" spaces automatically.
[0029] The algorithms calculates the pixels that will require filling. This is done through intersection and subtraction calculations on the masks or alpha channels of the layers above the current layer.
[0030] The software then uses a scanline algorithm, only filling one pixel per intersection per scanline until the area is filled. The color of each pixel is based on a weighted average of any surrounding pixels that already have color either from the original image or from earlier passes of this algorithm. The weight of each pixel is adjusted after each pass so that it relies evenly on top/bottom and right/left pixels for the first passes, and relies more heavily on the right/left pixels as the distance to original pixels increases.
[0031] For each layer, the user selects the depth by entering the value into the GUI edit field provided. Starting with the foreground, the software then assigns layer order by taking the layers in the order in which they were selected. The software then accurately generates the frames necessary for the interphasing step. See FIG. 1.1- 1.4.
[0032] Can be used when multiple images are input to produce three-dimensions from multiple image cameras, the case of different or sequential images, flip, morph, zoom or action images are produced.
[0033] To effect the conversion progress described above, the software package of this invention is controlled by the same optical principles used in three-dimensional photography. Such principles are described, for example in US Patent Nos . 3,852,787; 3,895,867; and 3,960,563; each of the references being hereby incorporated by reference herein in the entirety.
[0034] It should be understood that the particular embodiments shown in the drawings and described within this specification are for purposes of example and should not be construed to limit the invention which will be described in the claims below.

Claims

What is claimed is: 1. A method for converting a two-dimensional image file into a three dimensional image file, comprising steps of: receiving a two-dimensional pixel image data file into a data storage, the two-dimensional image file having pixel data reflecting a plurality of objects,- retrieving the two-dimensional pixel image data file from said storage; displaying an image corresponding to the image file on a video display; receiving a segmentation command indicating an initial border for at least one of the objects; displaying a border for the at least one object based on said segmentation command; segmenting the image file into a plurality of images based on the segmentation command,- in-painting at least one of the plurality of images to fill a portion at least one of said plurality of images with fill pixels, said fill pixels generated based on pixels within an object represented by the at least one image file; receiving a first depth data and a second depth data for at least a first and a second of the plurality images ; generating at least two phase-shifted images for each of the at least first and second of the plurality of images, the two phase-shifted images having a parallax shift with respect to one another and with respect to at least one other of the plurality of images, the parallax shift being based on the depth data for that image; interphasing said phase-shifted images into a single interphased image file for output and printing.
2. A method according to claim 1 further comprising a step of receiving a border refinement command having parameters for generating a final border for at least one of the objects, the parameters being from a group consisting of one or more of color, noise frequency, and edge softness, and wherein said segmentation is performed at least on part on aid border refinement command.
PCT/US2001/028563 2000-09-14 2001-09-14 Method for automated two-dimensional and three-dimensional conversion WO2002027667A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
AU2001290838A AU2001290838A1 (en) 2000-09-14 2001-09-14 Method for automated two-dimensional and three-dimensional conversion
JP2002531371A JP2004510272A (en) 2000-09-14 2001-09-14 Automatic 2D and 3D conversion method
EP01970887A EP1323135A1 (en) 2000-09-14 2001-09-14 Method for automated two-dimensional and three-dimensional conversion

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US23241000P 2000-09-14 2000-09-14
US60/232,410 2000-09-14

Publications (1)

Publication Number Publication Date
WO2002027667A1 true WO2002027667A1 (en) 2002-04-04

Family

ID=22872982

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2001/028563 WO2002027667A1 (en) 2000-09-14 2001-09-14 Method for automated two-dimensional and three-dimensional conversion

Country Status (6)

Country Link
EP (1) EP1323135A1 (en)
JP (1) JP2004510272A (en)
CN (1) CN1524249A (en)
AU (1) AU2001290838A1 (en)
RU (1) RU2003110175A (en)
WO (1) WO2002027667A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012014009A1 (en) * 2010-07-26 2012-02-02 City University Of Hong Kong Method for generating multi-view images from single image
CN102469318A (en) * 2010-11-04 2012-05-23 深圳Tcl新技术有限公司 Method for converting two-dimensional image into three-dimensional image
TWI463434B (en) * 2011-01-28 2014-12-01 Chunghwa Picture Tubes Ltd Image processing method for forming three-dimensional image from two-dimensional image
US9305398B2 (en) 2010-10-08 2016-04-05 City University Of Hong Kong Methods for creating and displaying two and three dimensional images on a digital canvas
US9432651B2 (en) 2008-07-24 2016-08-30 Koninklijke Philips N.V. Versatile 3-D picture format
US10133283B2 (en) 2012-07-26 2018-11-20 Honeywell International Inc. HVAC controller with wireless network based occupancy detection and control
US10928087B2 (en) 2012-07-26 2021-02-23 Ademco Inc. Method of associating an HVAC controller with an external web service
EP2306744B1 (en) * 2009-09-30 2021-05-19 Disney Enterprises, Inc. Method and system for utilizing pre-existing image layers of a two-dimensional image to create a stereoscopic image

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
ATE395670T1 (en) * 2004-11-16 2008-05-15 Koninkl Philips Electronics Nv IMAGE PLAYBACK BASED ON IMAGE SEGMENTATION
JP4463215B2 (en) * 2006-01-30 2010-05-19 日本電気株式会社 Three-dimensional processing apparatus and three-dimensional information terminal
JP5011316B2 (en) * 2006-02-27 2012-08-29 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Rendering the output image
KR100753536B1 (en) * 2006-05-04 2007-08-30 주식회사 아이너스기술 Method for detecting 2 dimension sketch data of source model data for 3 dimension reverse modeling
US7719531B2 (en) * 2006-05-05 2010-05-18 Microsoft Corporation Editing text within a three-dimensional graphic
CN101206767B (en) * 2006-12-22 2010-12-15 财团法人资讯工业策进会 Image conversion device and method for transforming flat picture to three-dimensional effect image
CA2686714C (en) * 2007-06-01 2016-11-01 Exxonmobil Upstream Research Company Generation of constrained voronoi grid in a plane
CN101350016B (en) * 2007-07-20 2010-11-24 富士通株式会社 Device and method for searching three-dimensional model
CN101271578B (en) * 2008-04-10 2010-06-02 清华大学 Depth sequence generation method of technology for converting plane video into stereo video
CN101593349B (en) * 2009-06-26 2012-06-13 福州华映视讯有限公司 Method for converting two-dimensional image into three-dimensional image
CN101860767B (en) * 2010-05-18 2011-10-19 南京大学 Lattice-based three-dimensional moving image display method and realization device thereof
CN102469323B (en) * 2010-11-18 2014-02-19 深圳Tcl新技术有限公司 Method for converting 2D (Two Dimensional) image to 3D (Three Dimensional) image
CN105117021A (en) * 2015-09-24 2015-12-02 深圳东方酷音信息技术有限公司 Virtual reality content generation method and playing device
CN105631938B (en) * 2015-12-29 2019-12-24 联想(北京)有限公司 Image processing method and electronic equipment

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4925294A (en) * 1986-12-17 1990-05-15 Geshwind David M Method to convert two dimensional motion pictures for three-dimensional systems

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4925294A (en) * 1986-12-17 1990-05-15 Geshwind David M Method to convert two dimensional motion pictures for three-dimensional systems

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9432651B2 (en) 2008-07-24 2016-08-30 Koninklijke Philips N.V. Versatile 3-D picture format
US10567728B2 (en) 2008-07-24 2020-02-18 Koninklijke Philips N.V. Versatile 3-D picture format
EP2306744B1 (en) * 2009-09-30 2021-05-19 Disney Enterprises, Inc. Method and system for utilizing pre-existing image layers of a two-dimensional image to create a stereoscopic image
WO2012014009A1 (en) * 2010-07-26 2012-02-02 City University Of Hong Kong Method for generating multi-view images from single image
US9305398B2 (en) 2010-10-08 2016-04-05 City University Of Hong Kong Methods for creating and displaying two and three dimensional images on a digital canvas
CN102469318A (en) * 2010-11-04 2012-05-23 深圳Tcl新技术有限公司 Method for converting two-dimensional image into three-dimensional image
TWI463434B (en) * 2011-01-28 2014-12-01 Chunghwa Picture Tubes Ltd Image processing method for forming three-dimensional image from two-dimensional image
US10133283B2 (en) 2012-07-26 2018-11-20 Honeywell International Inc. HVAC controller with wireless network based occupancy detection and control
US10613555B2 (en) 2012-07-26 2020-04-07 Ademco Inc. HVAC controller with wireless network based occupancy detection and control
US10928087B2 (en) 2012-07-26 2021-02-23 Ademco Inc. Method of associating an HVAC controller with an external web service
US11493224B2 (en) 2012-07-26 2022-11-08 Ademco Inc. Method of associating an HVAC controller with an external web service

Also Published As

Publication number Publication date
CN1524249A (en) 2004-08-25
EP1323135A1 (en) 2003-07-02
RU2003110175A (en) 2004-08-27
AU2001290838A1 (en) 2002-04-08
JP2004510272A (en) 2004-04-02

Similar Documents

Publication Publication Date Title
WO2002027667A1 (en) Method for automated two-dimensional and three-dimensional conversion
US4925294A (en) Method to convert two dimensional motion pictures for three-dimensional systems
JP7283513B2 (en) VIDEO DISPLAY DEVICE, VIDEO PROJECTION DEVICE, THEIR METHOD AND PROGRAM
JP4644669B2 (en) Multi-view image generation
US5311329A (en) Digital filtering for lenticular printing
KR100445619B1 (en) Device and method for converting two-dimensional video into three-dimensional video
JP2010154422A (en) Image processor
EP2323416A2 (en) Stereoscopic editing for video production, post-production and display adaptation
CN104395931A (en) Generation of a depth map for an image
CN104112275B (en) A kind of method and device for generating viewpoint
EP2225725A2 (en) Segmentation of image data
JP2009282979A (en) Image processor and image processing method
WO2003005299A2 (en) Image segmentation by means of temporal parallax difference induction
US20130343638A1 (en) Pulling keys from color segmented images
KR20110134142A (en) Method and apparatus for transforming stereoscopic image by using depth map information
JP2002123842A (en) Device for generating stereoscopic image, and medium for recording information
US11398007B2 (en) Video generation device, video generation method, program, and data structure
JPH11509998A (en) Method for converting an image into a stereoscopic image, and an image and a series of images obtained by the method
JPH09504131A (en) Image processing system for handling depth information
US6828973B2 (en) Method and system for 3-D object modeling
JP3091644B2 (en) 3D image conversion method for 2D images
JP2006211383A (en) Stereoscopic image processing apparatus, stereoscopic image display apparatus, and stereoscopic image generating method
JP5545059B2 (en) Moving image processing method, moving image processing apparatus, and moving image processing program
WO2012096065A1 (en) Parallax image display device and parallax image display method
KR101626679B1 (en) Method for generating stereoscopic image from 2D image and for medium recording the same

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2002531371

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 305/MUMNP/2003

Country of ref document: IN

WWE Wipo information: entry into national phase

Ref document number: 2001970887

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2003110175

Country of ref document: RU

Kind code of ref document: A

Ref country code: RU

Ref document number: RU A

WWE Wipo information: entry into national phase

Ref document number: 018188818

Country of ref document: CN

WWP Wipo information: published in national office

Ref document number: 2001970887

Country of ref document: EP

REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

WWW Wipo information: withdrawn in national office

Ref document number: 2001970887

Country of ref document: EP