CA2538162C - High speed multiple line three-dimensional digitization - Google Patents
High speed multiple line three-dimensional digitization Download PDFInfo
- Publication number
- CA2538162C CA2538162C CA002538162A CA2538162A CA2538162C CA 2538162 C CA2538162 C CA 2538162C CA 002538162 A CA002538162 A CA 002538162A CA 2538162 A CA2538162 A CA 2538162A CA 2538162 C CA2538162 C CA 2538162C
- Authority
- CA
- Canada
- Prior art keywords
- distinguishable
- lines
- curvilinear
- curvilinear lines
- line
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
- G01B11/25—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
- G01B11/2518—Projection by scanning of the object
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61C—DENTISTRY; APPARATUS OR METHODS FOR ORAL OR DENTAL HYGIENE
- A61C9/00—Impression cups, i.e. impression trays; Impression methods
- A61C9/004—Means or methods for taking digitized impressions
- A61C9/0046—Data acquisition means or methods
- A61C9/0053—Optical means or methods, e.g. scanning the teeth by a laser or light beam
- A61C9/006—Optical means or methods, e.g. scanning the teeth by a laser or light beam projecting one or more stripes or patterns on the teeth
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61C—DENTISTRY; APPARATUS OR METHODS FOR ORAL OR DENTAL HYGIENE
- A61C9/00—Impression cups, i.e. impression trays; Impression methods
- A61C9/004—Means or methods for taking digitized impressions
Abstract
A system provides high-speed multiple line digitization for three-dimensional imaging of a physical object. A full frame of three-dimensional data may be acquired in the same order as the frame rate of a digital camera.
Description
HIGH SPEED MULTIPLE LINE
TFIREE-DIMENSIONAL DIGITIZATION
BACKGROUND OF THE INVENTION
1. Technical Field
TFIREE-DIMENSIONAL DIGITIZATION
BACKGROUND OF THE INVENTION
1. Technical Field
[0002] This invention generally relates to three-dimensional imaging of a physical object, and in particular. to a high-speed multi-line triangulation for three-dimensional digitization of a physical object.
2. Related Art =[0003] Iinagiag techniques provide a three-dimensional visualization of a physical object- on a video terminal or monitor. The three-dimensional visualization may illustrate surface characteristics of the physical object.
Data associated with the surface characteristics are : generated and processed by a processor to generate the three-dimensional.visualization.
[0004] Data associated with the surface characteristics are generated by.
capturing images of the object from various perspectives. The perspectives are mapped or combined to produce a set of data points that represent the various 7 - . - . - . ' -surfaces of the object. The data points are processed to generate the three-dimensional visual display of the object. The data points also may be processed to I
represent the object in a dimensionally correct manner in the computer.
However, the time to generate the data points is longer than the display rate for the digital camera.
[0005] Imaging systems that use a triangulation technique emanate a single point or a single line on the object to determine relative surface characteristics of the object. Multiple line systems are limited by the maximum number of simultaneous lines that may image the object and require a large number of images to obtain a final image of the object.
[0006] A Moire technique may use multiple lines to compute a relative height map of the surface characteristics. Each point has a known or predetermined relative relationship to a neighboring point on a neighboring line.
A sinusoidal variation of the lines provides a trigonometric solution to estimate the relative relationships or equivalently extracting the phase. Such technique requires either multiple images or substantial processing time to provide a three-dimensional image.
[0007] Accordingly, there is a need for a high-speed three dimensional imaging system that minimizes the number of images and amount of computation to provide a three-dimensional image.
SUMMARY
[0008] By way of introduction only, a high speed multiple line three-dimensional digitization may include imaging a physical object to provide a visualization or virtualization of the object that may be viewed and manipulated by using a processor. The high speed multiple line three-dimensional digitization may be achieved by one or more apparatuses, devices, systems, methods, and/or processes.
[0009] In an embodiment, the high-speed multiple-line digitization generates a full frame of three-dimensional data of a surface of a physical object that is acquired in substantially the same order as a frame rate of a camera used to acquire or capture the three-dimensional image. For example, a camera used to capture a three-dimensional image has a frame rate of N frames per second. A
full frame of three-dimensional data is obtained in a time frame of m/N seconds where m different patterns are projected. For example, where two patterns are projected, a full frame of three-dimensional data is obtained in a time frame f 2/N
seconds.
The full frame of three-dimensional data includes multiple data points represented by multiple floating point numbers.
[0010] The foregoing summary is provided only by way of introduction.
The features and advantages of the high speed multiple line three-dimensional digitization may be realized and obtained by means of the instrumentalities and combinations particularly pointed out in the claims. Nothing in this section should be taken as a limitation on the claims, which define the scope of the invention.
Additional features and advantages of the present invention will be set forth in the description that follows, and in part will be obvious from the description, or may be learned by practice of the present invention.
2. Related Art =[0003] Iinagiag techniques provide a three-dimensional visualization of a physical object- on a video terminal or monitor. The three-dimensional visualization may illustrate surface characteristics of the physical object.
Data associated with the surface characteristics are : generated and processed by a processor to generate the three-dimensional.visualization.
[0004] Data associated with the surface characteristics are generated by.
capturing images of the object from various perspectives. The perspectives are mapped or combined to produce a set of data points that represent the various 7 - . - . - . ' -surfaces of the object. The data points are processed to generate the three-dimensional visual display of the object. The data points also may be processed to I
represent the object in a dimensionally correct manner in the computer.
However, the time to generate the data points is longer than the display rate for the digital camera.
[0005] Imaging systems that use a triangulation technique emanate a single point or a single line on the object to determine relative surface characteristics of the object. Multiple line systems are limited by the maximum number of simultaneous lines that may image the object and require a large number of images to obtain a final image of the object.
[0006] A Moire technique may use multiple lines to compute a relative height map of the surface characteristics. Each point has a known or predetermined relative relationship to a neighboring point on a neighboring line.
A sinusoidal variation of the lines provides a trigonometric solution to estimate the relative relationships or equivalently extracting the phase. Such technique requires either multiple images or substantial processing time to provide a three-dimensional image.
[0007] Accordingly, there is a need for a high-speed three dimensional imaging system that minimizes the number of images and amount of computation to provide a three-dimensional image.
SUMMARY
[0008] By way of introduction only, a high speed multiple line three-dimensional digitization may include imaging a physical object to provide a visualization or virtualization of the object that may be viewed and manipulated by using a processor. The high speed multiple line three-dimensional digitization may be achieved by one or more apparatuses, devices, systems, methods, and/or processes.
[0009] In an embodiment, the high-speed multiple-line digitization generates a full frame of three-dimensional data of a surface of a physical object that is acquired in substantially the same order as a frame rate of a camera used to acquire or capture the three-dimensional image. For example, a camera used to capture a three-dimensional image has a frame rate of N frames per second. A
full frame of three-dimensional data is obtained in a time frame of m/N seconds where m different patterns are projected. For example, where two patterns are projected, a full frame of three-dimensional data is obtained in a time frame f 2/N
seconds.
The full frame of three-dimensional data includes multiple data points represented by multiple floating point numbers.
[0010] The foregoing summary is provided only by way of introduction.
The features and advantages of the high speed multiple line three-dimensional digitization may be realized and obtained by means of the instrumentalities and combinations particularly pointed out in the claims. Nothing in this section should be taken as a limitation on the claims, which define the scope of the invention.
Additional features and advantages of the present invention will be set forth in the description that follows, and in part will be obvious from the description, or may be learned by practice of the present invention.
3 BRIEF DESCRIPTION OF THE DRAWINGS
[0011] The invention can be better understood with reference to the following drawings and description. The components in the figures are not necessarily to scale, emphasis instead being placed upon illustrating principles of the invention. Moreover, in the figures, like referenced numerals designate corresponding parts throughout the different views.
[0012] Figure 1 illustrates a high-speed multi-line three-dimensional digitization system having a camera fixed relative to a line projector.
[0013] Figure 2 illustrates a coordinate view of the system of Figure 1.
[0014] Figure 3 illustrates a line pattern incident on a surface of an object.
[0015] Figure 4 illustrates an approximate line pattern estimated from a first pattern frame scan for comparison with an actual line pattern for a subsequent frame as observed by the camera of Figure 1.
[0016] Figure 5 illustrates non-rectangular regions of distinguishability.
[0017] Figure 6 illustrates an embodiment where multiple light sources are derived from a single laser source switched with a liquid crystal cell.
[0018] Figure 7 illustrates an embodiment for multiple laser sources.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
[0019] Hereinafter exemplary embodiments are discussed with reference to accompanied figures.
[0020] Figure 1 illustrates a digitization system 100. The digitization system 100 includes a camera 106 and a line pattern projector or illuminator 104.
[0011] The invention can be better understood with reference to the following drawings and description. The components in the figures are not necessarily to scale, emphasis instead being placed upon illustrating principles of the invention. Moreover, in the figures, like referenced numerals designate corresponding parts throughout the different views.
[0012] Figure 1 illustrates a high-speed multi-line three-dimensional digitization system having a camera fixed relative to a line projector.
[0013] Figure 2 illustrates a coordinate view of the system of Figure 1.
[0014] Figure 3 illustrates a line pattern incident on a surface of an object.
[0015] Figure 4 illustrates an approximate line pattern estimated from a first pattern frame scan for comparison with an actual line pattern for a subsequent frame as observed by the camera of Figure 1.
[0016] Figure 5 illustrates non-rectangular regions of distinguishability.
[0017] Figure 6 illustrates an embodiment where multiple light sources are derived from a single laser source switched with a liquid crystal cell.
[0018] Figure 7 illustrates an embodiment for multiple laser sources.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
[0019] Hereinafter exemplary embodiments are discussed with reference to accompanied figures.
[0020] Figure 1 illustrates a digitization system 100. The digitization system 100 includes a camera 106 and a line pattern projector or illuminator 104.
4 The camera 106 is fixed relative to the line pattern projector 104. The projector 104 illuminates a portion of a surface 108 of object 102 with a light pattern.
The object may be any physical object capable of being imaged. In an embodiment, the object may be dentition or dental items including molds, castings, dentition, prepared dentition and the like.-[0021] Light reflected from the surface 108 is captured by the camera 106.
Based on the reflected light pattern, three-dimensional data representative of the illuminated surface 108 may be generated. The three-dimensional data may be processed to generate a three-dimensional image of the illuminated surface 108.
The camera 106 may be characterized by a local coordinate system XY, and the projector 104 characterized by a local coordinate system X'Y'.
[0022] Referring to Figure 2, pattern projector 204 projects a pattern during a capture or read period. The pattern may be considered to be an assembly of multiple points. The number of points may be finite or substantially infmite.
The size of the points may be finite or infinitesimal. An example of such a pattern is a pattern consisting of multiple lines. The pattern may be structured white light.
[0023] In an embodiment, camera 206 is a high-speed camera that images general patterns or multiple line patterns. The camera 206 may also capture multiple line patterns during a read period. The relationship shown in Figure refers to a single point in such a line pattern. A triangulation axis R may be defined as passing through an intersection of an axial ray from camera 206 and an axial ray of projector 204. The axis R may be perpendicular to an axial ray from camera 206 and an axial ray of projector 204. The triangulation axis R also may.
be substantially parallel to Y and Y'. A minimum angle 0 between a valid ray between the projector 204 relative to a valid axial ray of the camera 206 is non-zero.
[0024] A line projected by projector 204 represents a connected series of points or curvilinear segments where a normal vector n at any point along the curve obeys the following equation or rule:
In *RI ~ V-2 (1) According to Equation (1), the angle between a point on the curve and the triangulation axis R is greater than or equal to about 45 degrees. The line may have a cross-sectional intensity characterized by a function that is independent of Equation 1. The cross-sectional intensity may have a sinusoidal variation, a Gaussian profile, or any other function for cross-sectional intensity.
[0025] The local coordinate system XY of the camera 206 may be further characterized by a coordinate system XYZ, where the XY coordinate system defined by the camera include axis Z, which is substantially perpendicular to both the X-axis and the Y-axis. The axis Z includes a range of values for Z based on optics limitations. The values for Z may be based on distances dl and d2 such that di :5 z5 dz. A single point from a projected line incident on a plane perpendicular to Z will appear to be displaced in the X direction by AX. Based on a triangulation angle, the following condition exists:
~ Ax (2) Tan B
[0026] In a projected line pattern having multiple lines LI-L,õ a given line L; may be characterized by a unique function 9(x). For a given line L;, the location of line L; with respect to the coordinate system XYZ of the camera for various values of z where d, S z<_ d2 may be determined through calibration or similar process.
[0027] For an observed line Lc, a closest calibrated line position may be selected, and the x and z coordinates (x,,z,;) of the calibrated line determined. The camera 206 may observe multiple lines during projected on an object 102. For each observed point on the line, as captured or observed by the camera 206, the XY coordinates of that surface point may be directly observed as (xobse,ved,yobserved)=
A point Zobserved may be determined by observing the displaced Ax (where Ax =
xobserved'xc), to compute Az. The z coordinate may then be computed as:
Zobsen,ed - Zc + A-z' (3) The maximum displacement for any line in the volume may be determined by:
Ax = (di - dz )Tari 0 (4) [0028] A maximum nuinber of simultaneously distinguishable lines flm.
may be determined as:
nmax - X max (4) - (d, - dZ )Tan Bmax [0029] The maximum number of simultaneously distinguishable lines nm.
increases with a decreasing depth of field di - dz . The maximum number of simultaneously distinguishable lines n,,,ax also increases with as Bm-decreases.
The accuracy of the determination also inay also decrease with smaller max values. Also, decreasing a depth of field may result in a less useful volume for digitizing.
[00301 Figure 3 illustrates a line pattern having multiple lines Li-L"
projected toward an object 302. Each line LI-Lõ represents a connected series of points or curvilinear segments where a normal vector n at any point along the curve obeys equation 1 above. The multiple lines LI-Lõ are projected toward and incident onto a surface 308 of the object 302.
[0031] Multiple patterns of lines LI-Lt, may be projected toward the object 302 during a capture period. The light patterns may be referred to as A; where i =
1, 2, ....x, where the first light pattern having LI-Lõ lines is referred to as A1 and subsequent line patterns are referred to as A2 to A,. The number of lines n in pattern A; may be selected so that n< n,,,,.
[0032] According to Equation (4), each line in pattern A, incident on the surface 308 may be uniquely labeled or identified. For each line pattern A1, the x, y and z coordinates may be determined for each point on the line using the above equations. For each line L;, data-points representative of characteristics of the surface 308 along the line L; may be generated. From the data points, a three-dimensional representation of the surface 302 along the line L; is formed.
From all the lines of pattern Ai, an approximation of the surface of the object being digitized may be determined.
[0033] For the subsequent patterns A;, where i= 2, ....x, let n, represent the number of lines for the pattern A. For i< j the condition n, _< n, holds.
Also, n; >
Yin=ax for each i. Because equation (4) no longer holds, labeling or identifying lines for Ai may be resolved during a prior calibration step.
[0034] In a calibration step, each line in A; is characterized on a flat plane for different Z values. Based on the characterization, and an approximation surface, the approximate locations of each labeled line in A; is estimated by intersecting a known light plane corresponding with each labeled line with the approximation surface 308. The estimation may be compared to the observed line pattern for A; incident on the surface 302, and observed lines accordingly labeled.
[00351 Figure 4 illustrates an approximate line pattern estimated from a first pattern frame scan 410 coinpared with an actual line pattern 412 for a subsequent frame as observed by camera 206. By choosing closest curves, a unique labeling of the multiple lines Lt-Lõ is )obtained. A new approximation surface is thus obtained by repeated application of equation (4) on each labeled line. This may be repeated using a new and enhanced approximation surface of the surface and a higher density line pattern.
[0036] Figure 5 illustrates non-rectangular regions of distinguishability. In an embodiment, non-rectangular regions of distinguishability may be defined as areas between adjacent projection lines Ll - L,,. For example, a non-rectangular region of distinguishability may be defined as the region between a first line LI of the multiple line pattern and a second line L2 of the light pattern. For each line that may be projected onto a planar surface placed at z values between dl and d2, the region of distinguishability defines the smallest envelope that always includes that line as imaged by the imaging system. Other lines L; will have separate regions of distinguishability. Therefore, in the exemplary embodiment, each line may be projected to a discrete area where substantially no overlap exists between adjacent areas.
[0037] Figure 5 illustrates an example of a pattern where three lines are being projected, with non-overlapping regions of distinguishability. By allowing non-rectangular regions of distinguishability, the limitations of equation (4) may be minimized or eliminated altogether by allowing non-rectangular regions for each line, where the non-rectangular regions may be compressed without substantial overlap or ambiguity. The number of simultaneous projected lines may be increased by allowing the distinguishable regions around each line to follow a shape of the line instead of a defined rectangular area. Accordingly, the maximum number of simultaneous lines may be increased.
[0038] An embodiment for a projector for a high speed multiple line three-dimensional digitization system may include a modulated laser source having a two-axis orthogonal mirror scanner. The scanner may have a single two axis mirror, two orthogonally mounted single axis mirrors, or the equivalents. The scanner may project a two-dimensional light pattern having multiple lines Li-Lõ
through optics toward a surface of an object. The light pattern illuminates the surface. Light reflected from the surface may be captured by a camera. The patterns incident on the object may be viewed through additional optics, a CCD, CMOS digital camera, or similar device. The line patterns are analyzed and converted to three-dimensional coordinates representative of the illuminated surface.
100391 Referring to Figure 6, an embodiment having two laser sources 620, 622 is shown. The two laser sources 620, 622 each project a laser beam. The laser beams pass through a focusing lens 624 and a line lens 626 that transforms the single point from each laser source 620, 622 into a line. Each line then passes through two different diffraction gratings 628 that split the line into multiple substantially parallel lines: The multiple minimal patterns may be produced without substantial moving parts.
[0040] Referring to Figure 7, an embodiment having a single laser source 720 is shown. The laser source 720 may be switched using a liquid crystal (LC).
ceII 734. The laser light beam from the single laser source 720 passes through a wave plate 732 to polarize the light. The wave plate 732 may be a%z wave plate.
The laser light is polarized based on the LC cell 734, in a particular direction and/or orientation, where the direction may be switched by switching-the LC
cell 734. Two possible paths for the light may pass through the beam splitter 736 so that the single laser beam may be split to two paths. The laser point may be processed as described.
[0041] In an embodiment for a high speed multiple line three-dimensional digitization system, a scanner and camera may be configured as described according to co-pending US Patent No. 7,142,312 entitled LASER DIGITIZER
SYSTEM FOR DENTAL APPLICATIONS, filed on December 30, 2003.
The scanner may be a modulated laser source, coupled to a two axis orthogonal mirror scanner. An embodiment for a high speed multiple line three-dimensional digitization system may include a modulated laser source, couple4 to a two axis orthogonal mirror scanner. The scanner may have a single two axis mirror, two orthogonally mounted single axis mirrors, or the equivalents. By varying the rotation of the mirror(s), and by modulating the laser beam, a two-dimensional pattern may be traced. The pattern may be projected through optics onto the physical object, and the patterns incident on the object viewed through additional optics, a CCD, CMOS digital camera, or similar device. The line patterns are analyzed and converted to three-dimensional coordinates for the surface.
[0042] It is intended in the appended claims to cover all such changes and modifications which fall within the true spirit and scope of the invention.
Therefore, the invention is not limited to the specific details, representative embodiments, and illustrated examples in this description. Accordingly, the invention is not to be restricted except in light as necessitated by the accompanying claims and their equivalents.
The object may be any physical object capable of being imaged. In an embodiment, the object may be dentition or dental items including molds, castings, dentition, prepared dentition and the like.-[0021] Light reflected from the surface 108 is captured by the camera 106.
Based on the reflected light pattern, three-dimensional data representative of the illuminated surface 108 may be generated. The three-dimensional data may be processed to generate a three-dimensional image of the illuminated surface 108.
The camera 106 may be characterized by a local coordinate system XY, and the projector 104 characterized by a local coordinate system X'Y'.
[0022] Referring to Figure 2, pattern projector 204 projects a pattern during a capture or read period. The pattern may be considered to be an assembly of multiple points. The number of points may be finite or substantially infmite.
The size of the points may be finite or infinitesimal. An example of such a pattern is a pattern consisting of multiple lines. The pattern may be structured white light.
[0023] In an embodiment, camera 206 is a high-speed camera that images general patterns or multiple line patterns. The camera 206 may also capture multiple line patterns during a read period. The relationship shown in Figure refers to a single point in such a line pattern. A triangulation axis R may be defined as passing through an intersection of an axial ray from camera 206 and an axial ray of projector 204. The axis R may be perpendicular to an axial ray from camera 206 and an axial ray of projector 204. The triangulation axis R also may.
be substantially parallel to Y and Y'. A minimum angle 0 between a valid ray between the projector 204 relative to a valid axial ray of the camera 206 is non-zero.
[0024] A line projected by projector 204 represents a connected series of points or curvilinear segments where a normal vector n at any point along the curve obeys the following equation or rule:
In *RI ~ V-2 (1) According to Equation (1), the angle between a point on the curve and the triangulation axis R is greater than or equal to about 45 degrees. The line may have a cross-sectional intensity characterized by a function that is independent of Equation 1. The cross-sectional intensity may have a sinusoidal variation, a Gaussian profile, or any other function for cross-sectional intensity.
[0025] The local coordinate system XY of the camera 206 may be further characterized by a coordinate system XYZ, where the XY coordinate system defined by the camera include axis Z, which is substantially perpendicular to both the X-axis and the Y-axis. The axis Z includes a range of values for Z based on optics limitations. The values for Z may be based on distances dl and d2 such that di :5 z5 dz. A single point from a projected line incident on a plane perpendicular to Z will appear to be displaced in the X direction by AX. Based on a triangulation angle, the following condition exists:
~ Ax (2) Tan B
[0026] In a projected line pattern having multiple lines LI-L,õ a given line L; may be characterized by a unique function 9(x). For a given line L;, the location of line L; with respect to the coordinate system XYZ of the camera for various values of z where d, S z<_ d2 may be determined through calibration or similar process.
[0027] For an observed line Lc, a closest calibrated line position may be selected, and the x and z coordinates (x,,z,;) of the calibrated line determined. The camera 206 may observe multiple lines during projected on an object 102. For each observed point on the line, as captured or observed by the camera 206, the XY coordinates of that surface point may be directly observed as (xobse,ved,yobserved)=
A point Zobserved may be determined by observing the displaced Ax (where Ax =
xobserved'xc), to compute Az. The z coordinate may then be computed as:
Zobsen,ed - Zc + A-z' (3) The maximum displacement for any line in the volume may be determined by:
Ax = (di - dz )Tari 0 (4) [0028] A maximum nuinber of simultaneously distinguishable lines flm.
may be determined as:
nmax - X max (4) - (d, - dZ )Tan Bmax [0029] The maximum number of simultaneously distinguishable lines nm.
increases with a decreasing depth of field di - dz . The maximum number of simultaneously distinguishable lines n,,,ax also increases with as Bm-decreases.
The accuracy of the determination also inay also decrease with smaller max values. Also, decreasing a depth of field may result in a less useful volume for digitizing.
[00301 Figure 3 illustrates a line pattern having multiple lines Li-L"
projected toward an object 302. Each line LI-Lõ represents a connected series of points or curvilinear segments where a normal vector n at any point along the curve obeys equation 1 above. The multiple lines LI-Lõ are projected toward and incident onto a surface 308 of the object 302.
[0031] Multiple patterns of lines LI-Lt, may be projected toward the object 302 during a capture period. The light patterns may be referred to as A; where i =
1, 2, ....x, where the first light pattern having LI-Lõ lines is referred to as A1 and subsequent line patterns are referred to as A2 to A,. The number of lines n in pattern A; may be selected so that n< n,,,,.
[0032] According to Equation (4), each line in pattern A, incident on the surface 308 may be uniquely labeled or identified. For each line pattern A1, the x, y and z coordinates may be determined for each point on the line using the above equations. For each line L;, data-points representative of characteristics of the surface 308 along the line L; may be generated. From the data points, a three-dimensional representation of the surface 302 along the line L; is formed.
From all the lines of pattern Ai, an approximation of the surface of the object being digitized may be determined.
[0033] For the subsequent patterns A;, where i= 2, ....x, let n, represent the number of lines for the pattern A. For i< j the condition n, _< n, holds.
Also, n; >
Yin=ax for each i. Because equation (4) no longer holds, labeling or identifying lines for Ai may be resolved during a prior calibration step.
[0034] In a calibration step, each line in A; is characterized on a flat plane for different Z values. Based on the characterization, and an approximation surface, the approximate locations of each labeled line in A; is estimated by intersecting a known light plane corresponding with each labeled line with the approximation surface 308. The estimation may be compared to the observed line pattern for A; incident on the surface 302, and observed lines accordingly labeled.
[00351 Figure 4 illustrates an approximate line pattern estimated from a first pattern frame scan 410 coinpared with an actual line pattern 412 for a subsequent frame as observed by camera 206. By choosing closest curves, a unique labeling of the multiple lines Lt-Lõ is )obtained. A new approximation surface is thus obtained by repeated application of equation (4) on each labeled line. This may be repeated using a new and enhanced approximation surface of the surface and a higher density line pattern.
[0036] Figure 5 illustrates non-rectangular regions of distinguishability. In an embodiment, non-rectangular regions of distinguishability may be defined as areas between adjacent projection lines Ll - L,,. For example, a non-rectangular region of distinguishability may be defined as the region between a first line LI of the multiple line pattern and a second line L2 of the light pattern. For each line that may be projected onto a planar surface placed at z values between dl and d2, the region of distinguishability defines the smallest envelope that always includes that line as imaged by the imaging system. Other lines L; will have separate regions of distinguishability. Therefore, in the exemplary embodiment, each line may be projected to a discrete area where substantially no overlap exists between adjacent areas.
[0037] Figure 5 illustrates an example of a pattern where three lines are being projected, with non-overlapping regions of distinguishability. By allowing non-rectangular regions of distinguishability, the limitations of equation (4) may be minimized or eliminated altogether by allowing non-rectangular regions for each line, where the non-rectangular regions may be compressed without substantial overlap or ambiguity. The number of simultaneous projected lines may be increased by allowing the distinguishable regions around each line to follow a shape of the line instead of a defined rectangular area. Accordingly, the maximum number of simultaneous lines may be increased.
[0038] An embodiment for a projector for a high speed multiple line three-dimensional digitization system may include a modulated laser source having a two-axis orthogonal mirror scanner. The scanner may have a single two axis mirror, two orthogonally mounted single axis mirrors, or the equivalents. The scanner may project a two-dimensional light pattern having multiple lines Li-Lõ
through optics toward a surface of an object. The light pattern illuminates the surface. Light reflected from the surface may be captured by a camera. The patterns incident on the object may be viewed through additional optics, a CCD, CMOS digital camera, or similar device. The line patterns are analyzed and converted to three-dimensional coordinates representative of the illuminated surface.
100391 Referring to Figure 6, an embodiment having two laser sources 620, 622 is shown. The two laser sources 620, 622 each project a laser beam. The laser beams pass through a focusing lens 624 and a line lens 626 that transforms the single point from each laser source 620, 622 into a line. Each line then passes through two different diffraction gratings 628 that split the line into multiple substantially parallel lines: The multiple minimal patterns may be produced without substantial moving parts.
[0040] Referring to Figure 7, an embodiment having a single laser source 720 is shown. The laser source 720 may be switched using a liquid crystal (LC).
ceII 734. The laser light beam from the single laser source 720 passes through a wave plate 732 to polarize the light. The wave plate 732 may be a%z wave plate.
The laser light is polarized based on the LC cell 734, in a particular direction and/or orientation, where the direction may be switched by switching-the LC
cell 734. Two possible paths for the light may pass through the beam splitter 736 so that the single laser beam may be split to two paths. The laser point may be processed as described.
[0041] In an embodiment for a high speed multiple line three-dimensional digitization system, a scanner and camera may be configured as described according to co-pending US Patent No. 7,142,312 entitled LASER DIGITIZER
SYSTEM FOR DENTAL APPLICATIONS, filed on December 30, 2003.
The scanner may be a modulated laser source, coupled to a two axis orthogonal mirror scanner. An embodiment for a high speed multiple line three-dimensional digitization system may include a modulated laser source, couple4 to a two axis orthogonal mirror scanner. The scanner may have a single two axis mirror, two orthogonally mounted single axis mirrors, or the equivalents. By varying the rotation of the mirror(s), and by modulating the laser beam, a two-dimensional pattern may be traced. The pattern may be projected through optics onto the physical object, and the patterns incident on the object viewed through additional optics, a CCD, CMOS digital camera, or similar device. The line patterns are analyzed and converted to three-dimensional coordinates for the surface.
[0042] It is intended in the appended claims to cover all such changes and modifications which fall within the true spirit and scope of the invention.
Therefore, the invention is not limited to the specific details, representative embodiments, and illustrated examples in this description. Accordingly, the invention is not to be restricted except in light as necessitated by the accompanying claims and their equivalents.
Claims (22)
1. A method for digitizing a physical object, comprising:
(a) projecting a plurality of sequential light patterns onto a surface of an object, where each of the plurality of sequential patterns includes a plurality of distinguishable curvilinear lines, where a first pattern includes a first number of distinguishable curvilinear lines, and a subsequent pattern includes at least the same number of lines than the first number of distinguishable curvilinear lines;
(b) capturing light reflected from the surface;
(c) determining coordinates x, y, z of points on the surface according to the captured light; and (d) using the coordinates to generate a representation of the physical object.
(a) projecting a plurality of sequential light patterns onto a surface of an object, where each of the plurality of sequential patterns includes a plurality of distinguishable curvilinear lines, where a first pattern includes a first number of distinguishable curvilinear lines, and a subsequent pattern includes at least the same number of lines than the first number of distinguishable curvilinear lines;
(b) capturing light reflected from the surface;
(c) determining coordinates x, y, z of points on the surface according to the captured light; and (d) using the coordinates to generate a representation of the physical object.
2. The method of claim 1 where the plurality of sequential patterns comprises white light.
3. The method of claim 1 where each of the plurality of distinguishable curvilinear lines is characterized by an intensity profile being substantially constant over time.
4. The method of claim 1 where step (a) further comprises generating a light pattern according to tracing a laser point with a two axis mirror to generate the plurality of distinguishable curvilinear lines.
5. The method of claim 4 where each of plurality of distinguishable curvilinear lines is characterized by a modulated intensity pattern.
6. The method of claim 1 where step (a) further comprises generating a light pattern according to tracing a laser point with a plurality of orthogonally mounted single axis mirrors to generate the plurality of distinguishable curvilinear lines.
7. The method of claim 1 further comprising optically generating the plurality of distinguishable curvilinear lines.
8. The method of claim 7 where each of the plurality of distinguishable curvilinear lines is characterized by a modulated intensity pattern and a scan speed.
9. The method of claim 8 where the scan speed and modulation are variable.
10. The method of claim 1 further comprising optically projecting multiple laser points through a plurality of optical elements to generate a plurality of static patterns of illuminated lines.
11. The method of claim 10 further comprising a single laser source to generate the plurality of distinguishable curvilinear lines.
12. The method of claim 11 further comprising producing the plurality of distinguishable curvilinear lines using a mechanical switch.
13. The method of claim 12 where the mechanical switch comprises a mirror.
14. The method of claim 13 where the plurality of distinguishable curvilinear lines are produced using an electro-optical switch.
15. The method of claim 14 where the electro-optical switch comprises an LC
switch.
switch.
16. The method of claim 1 further comprising processing the coordinates in a computer system.
17. The method of claim 1 where the subsequent pattern includes a greater number of lines than the first number of distinguishable curvilinear lines.
18. The method of claim 1 further comprising displaying the coordinates as a visual representation of the surface.
19. A dental imaging system, comprising:
(a) means for projecting a plurality of sequential patterns toward a dental item, each of the plurality of sequential patterns including a plurality of distinguishable curvilinear lines, where a first pattern includes a first number of distinguishable curvilinear lines, and a subsequent pattern includes a second number of distinguishable curvilinear lines, wherein within a pattern each curvilinear line is associated with a region that is non-rectangular and that does not overlap with a region associated with an adjacent curvilinear line;
(b) means for capturing light reflected from a surface of the dental item;
(c) means for determining data points associated with characteristics of the surface according to the captured light reflected from the surface; and (d) means for displaying the a three-dimensional representation of the surface of the dental item according to the data points.
(a) means for projecting a plurality of sequential patterns toward a dental item, each of the plurality of sequential patterns including a plurality of distinguishable curvilinear lines, where a first pattern includes a first number of distinguishable curvilinear lines, and a subsequent pattern includes a second number of distinguishable curvilinear lines, wherein within a pattern each curvilinear line is associated with a region that is non-rectangular and that does not overlap with a region associated with an adjacent curvilinear line;
(b) means for capturing light reflected from a surface of the dental item;
(c) means for determining data points associated with characteristics of the surface according to the captured light reflected from the surface; and (d) means for displaying the a three-dimensional representation of the surface of the dental item according to the data points.
20. A dental digitizing system, comprising:
(a) a projector configured to illuminate a dental object during each of sequence of exposure periods with a light patterns including a plurality of distinguishable curvilinear lines, where a first light pattern includes a first number of distinguishable curvilinear lines, and a subsequent light pattern includes a second number of distinguishable curvilinear lines, wherein within a pattern each curvilinear line is associated with a region that is non-rectangular and that does not overlap with a region associated with an adjacent curvilinear line;
(b) a camera configured to capture an image according to light reflected during each exposure period;
(c) a controller configured to determine a set of points associated with characteristics of a surface of the object according to each image, and to compile each of the captured images to form an image of the object; and (d) a display for displaying the image of the object.
(a) a projector configured to illuminate a dental object during each of sequence of exposure periods with a light patterns including a plurality of distinguishable curvilinear lines, where a first light pattern includes a first number of distinguishable curvilinear lines, and a subsequent light pattern includes a second number of distinguishable curvilinear lines, wherein within a pattern each curvilinear line is associated with a region that is non-rectangular and that does not overlap with a region associated with an adjacent curvilinear line;
(b) a camera configured to capture an image according to light reflected during each exposure period;
(c) a controller configured to determine a set of points associated with characteristics of a surface of the object according to each image, and to compile each of the captured images to form an image of the object; and (d) a display for displaying the image of the object.
21. The dental digitizing of claim 20 where the projector comprises a laser source configured to optically generate the plurality of distinguishable curvilinear lines from a single point laser source.
22. The dental digitizing of claim 20 where the projector comprises a plurality of laser sources configured to generate the plurality of distinguishable curvilinear lines.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US50366603P | 2003-09-17 | 2003-09-17 | |
US60/503,666 | 2003-09-17 | ||
PCT/US2004/030453 WO2005027770A2 (en) | 2003-09-17 | 2004-09-17 | High speed multiple line three-dimensional digitization |
Publications (2)
Publication Number | Publication Date |
---|---|
CA2538162A1 CA2538162A1 (en) | 2005-03-31 |
CA2538162C true CA2538162C (en) | 2010-02-02 |
Family
ID=34375379
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA002538162A Active CA2538162C (en) | 2003-09-17 | 2004-09-17 | High speed multiple line three-dimensional digitization |
Country Status (6)
Country | Link |
---|---|
US (1) | US7342668B2 (en) |
EP (1) | EP1663046A4 (en) |
JP (1) | JP4913597B2 (en) |
AU (1) | AU2004273957B2 (en) |
CA (1) | CA2538162C (en) |
WO (1) | WO2005027770A2 (en) |
Families Citing this family (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
ATE555743T1 (en) | 2000-11-08 | 2012-05-15 | Straumann Inst Ag | (DENTAL) SURFACE CAPTURE AND CREATION |
AU2003252253A1 (en) | 2002-07-26 | 2004-02-16 | Olympus Optical Co., Ltd. | Image processing system |
AU2003252254A1 (en) * | 2002-07-26 | 2004-05-04 | Olympus Corporation | Image processing system |
EP1870667A3 (en) * | 2002-10-18 | 2008-03-19 | Aepsilon Rechteverwaltungs GmbH | Devices and methods for capturing surfaces and for producing denture elements |
JP5189287B2 (en) * | 2003-03-24 | 2013-04-24 | ディーフォーディー テクノロジーズ エルエルシー | Dental laser digitizer system |
CN101701849A (en) * | 2004-01-23 | 2010-05-05 | 奥林巴斯株式会社 | Image processing system and camera |
DE102006061134A1 (en) * | 2006-12-22 | 2008-06-26 | Aepsilon Rechteverwaltungs Gmbh | Process for the transport of dental prostheses |
US20090081617A1 (en) * | 2007-09-21 | 2009-03-26 | D4D Technologies, Llc | Display interface target positioning tool |
US8144954B2 (en) * | 2007-11-08 | 2012-03-27 | D4D Technologies, Llc | Lighting compensated dynamic texture mapping of 3-D models |
DE102007058590B4 (en) | 2007-12-04 | 2010-09-16 | Sirona Dental Systems Gmbh | Recording method for an image of a recording object and recording device |
US7821649B2 (en) * | 2008-03-05 | 2010-10-26 | Ge Inspection Technologies, Lp | Fringe projection system and method for a probe suitable for phase-shift analysis |
US8107083B2 (en) * | 2008-03-05 | 2012-01-31 | General Electric Company | System aspects for a probe system that utilizes structured-light |
US8422030B2 (en) * | 2008-03-05 | 2013-04-16 | General Electric Company | Fringe projection system with intensity modulating by columns of a plurality of grating elements |
DE102008015499C5 (en) * | 2008-03-25 | 2013-01-10 | Steinbichler Optotechnik Gmbh | Method and device for determining the 3D coordinates of an object |
IT1391808B1 (en) * | 2008-10-31 | 2012-01-27 | Sicam Srl | MALE DISPENSER FOR THE ASSEMBLY AND DISASSEMBLY OF VEHICLE WHEELS |
US8339616B2 (en) * | 2009-03-31 | 2012-12-25 | Micrometric Vision Technologies | Method and apparatus for high-speed unconstrained three-dimensional digitalization |
EP2272417B1 (en) * | 2009-07-10 | 2016-11-09 | GE Inspection Technologies, LP | Fringe projection system for a probe suitable for phase-shift analysis |
US8503539B2 (en) | 2010-02-26 | 2013-08-06 | Bao Tran | High definition personal computer (PC) cam |
US8134719B2 (en) * | 2010-03-19 | 2012-03-13 | Carestream Health, Inc. | 3-D imaging using telecentric defocus |
CN101853521B (en) * | 2010-04-22 | 2012-07-04 | 王少华 | Cultural relic rotation structured light three-dimensional digital modeling method |
US9157733B2 (en) * | 2010-09-10 | 2015-10-13 | Dimensional Photonics International, Inc. | Method of data acquisition for three-dimensional imaging |
US9349182B2 (en) * | 2011-11-10 | 2016-05-24 | Carestream Health, Inc. | 3D intraoral measurements using optical multiline method |
US9295532B2 (en) * | 2011-11-10 | 2016-03-29 | Carestream Health, Inc. | 3D intraoral measurements using optical multiline method |
DE102012220048B4 (en) | 2012-11-02 | 2018-09-20 | Sirona Dental Systems Gmbh | Calibration device and method for calibrating a dental camera |
CA2900268C (en) * | 2013-02-04 | 2021-04-06 | D4D Technologies, Llc | Intra-oral scanning device with illumination frames interspersed with image frames |
USD780182S1 (en) * | 2013-03-11 | 2017-02-28 | D4D Technologies, Llc | Handheld scanner |
ES2680587T3 (en) | 2014-08-28 | 2018-09-10 | Carestream Dental Technology Topco Limited | Intraoral 3-D measurements using a multi-line optical procedure |
WO2016099321A1 (en) * | 2014-12-19 | 2016-06-23 | Андрей Владимирович КЛИМОВ | Method for checking the linear dimensions of three-dimensional objects |
US11172824B2 (en) * | 2016-04-06 | 2021-11-16 | Carestream Dental Technology Topco Limited | Hybrid OCT and surface contour dental imaging |
PL3645964T3 (en) * | 2017-06-30 | 2023-12-18 | Dental Imaging Technologies Corporation | Surface mapping using an intraoral scanner with penetrating capabilities |
WO2019032923A2 (en) | 2017-08-10 | 2019-02-14 | D4D Technologies, Llc | Intra-oral scanning device |
Family Cites Families (117)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE3003435A1 (en) | 1980-01-31 | 1981-08-06 | Becker Dental-Labor Gmbh, 5100 Aachen | METHOD AND DEVICE FOR PRODUCING A CROWN PART |
US4575805A (en) * | 1980-12-24 | 1986-03-11 | Moermann Werner H | Method and apparatus for the fabrication of custom-shaped implants |
FR2499736B1 (en) * | 1981-02-12 | 1987-03-06 | Renault | DEVICE AND METHOD FOR LOCATING THREE-DIMENSIONAL BULK OBJECTS FOR CONTROLLING A GRIPPING TERMINAL |
DE3203937C2 (en) | 1982-02-05 | 1985-10-03 | Luc Dr. 4150 Krefeld Barrut | Method and device for machine restoration or correction of at least one tooth or for machine preparation of at least one tooth for a fixed prosthetic restoration and for machine production of the fixed prosthetic restoration |
FR2525103B1 (en) | 1982-04-14 | 1985-09-27 | Duret Francois | IMPRESSION TAKING DEVICE BY OPTICAL MEANS, PARTICULARLY FOR THE AUTOMATIC PRODUCTION OF PROSTHESES |
FR2536654B1 (en) | 1982-11-30 | 1987-01-09 | Duret Francois | METHOD FOR PRODUCING A DENTAL PROSTHESIS |
US4663720A (en) | 1984-02-21 | 1987-05-05 | Francois Duret | Method of and apparatus for making a prosthesis, especially a dental prosthesis |
CH665551A5 (en) | 1984-03-06 | 1988-05-31 | Werner Hans Dr Med De Moermann | BLANK FOR THE MANUFACTURE OF DENTAL TECHNOLOGY MOLDED PARTS. |
JPH0612252B2 (en) * | 1984-03-21 | 1994-02-16 | 友彦 芥田 | Automatic three-dimensional shape measurement method |
EP0163076B1 (en) | 1984-04-17 | 1991-11-13 | Kawasaki Jukogyo Kabushiki Kaisha | Apparatus for producing a three-dimensional copy of an object |
US4798534A (en) | 1984-08-03 | 1989-01-17 | Great Lakes Orthodontic Laboratories Inc. | Method of making a dental appliance |
CH663891A5 (en) | 1984-10-24 | 1988-01-29 | Marco Dr Sc Techn Brandestini | DEVICE FOR THE SHAPING PROCESSING OF A BLANK MADE OF DENTAL CERAMIC OR DENTAL COMPOSITE MATERIAL AND METHOD FOR THE OPERATION THEREOF. |
US4687326A (en) * | 1985-11-12 | 1987-08-18 | General Electric Company | Integrated range and luminance camera |
US4936862A (en) | 1986-05-30 | 1990-06-26 | Walker Peter S | Method of designing and manufacturing a human joint prosthesis |
CH672722A5 (en) * | 1986-06-24 | 1989-12-29 | Marco Brandestini | |
US4816920A (en) | 1986-11-18 | 1989-03-28 | General Scanning, Inc. | Planar surface scanning system |
FR2610821B1 (en) | 1987-02-13 | 1989-06-09 | Hennson Int | METHOD FOR TAKING MEDICAL IMPRESSION AND DEVICE FOR IMPLEMENTING SAME |
US4856991A (en) | 1987-05-05 | 1989-08-15 | Great Lakes Orthodontics, Ltd. | Orthodontic finishing positioner and method of construction |
US5186623A (en) | 1987-05-05 | 1993-02-16 | Great Lakes Orthodontics, Ltd. | Orthodontic finishing positioner and method of construction |
DE3723555C2 (en) | 1987-07-16 | 1994-08-11 | Steinbichler Hans | Process for the production of dentures |
US5372502A (en) | 1988-09-02 | 1994-12-13 | Kaltenbach & Voight Gmbh & Co. | Optical probe and method for the three-dimensional surveying of teeth |
US5055039A (en) | 1988-10-06 | 1991-10-08 | Great Lakes Orthodontics, Ltd. | Orthodontic positioner and methods of making and using same |
US4935635A (en) | 1988-12-09 | 1990-06-19 | Harra Dale G O | System for measuring objects in three dimensions |
US5011405A (en) | 1989-01-24 | 1991-04-30 | Dolphin Imaging Systems | Method for determining orthodontic bracket placement |
SE464908B (en) | 1989-03-23 | 1991-07-01 | Nobelpharma Ab | METHOD FOR MANUFACTURING ARTIFICIAL DENTAL CHRONICLES OF ONLINE TYPE OR INPUT |
JPH02268207A (en) * | 1989-04-10 | 1990-11-01 | Agency Of Ind Science & Technol | Shape recognizing apparatus |
US5151044A (en) | 1989-05-12 | 1992-09-29 | Rotsaert Henri L | Blanks for the manufacture of artificial teeth and crowns |
US4970032A (en) | 1989-05-12 | 1990-11-13 | Rotsaert Henri L | Processes for the manufacture of artificial teeth and crowns |
US5027281A (en) * | 1989-06-09 | 1991-06-25 | Regents Of The University Of Minnesota | Method and apparatus for scanning and recording of coordinates describing three dimensional objects of complex and unique geometry |
JPH0374310A (en) * | 1989-08-14 | 1991-03-28 | Mitsubishi Rayon Co Ltd | Adhesive composition for dental use |
JPH0723684Y2 (en) * | 1989-11-21 | 1995-05-31 | 横河電機株式会社 | Range finder |
US5368478A (en) | 1990-01-19 | 1994-11-29 | Ormco Corporation | Method for forming jigs for custom placement of orthodontic appliances on teeth |
US5533895A (en) | 1990-01-19 | 1996-07-09 | Ormco Corporation | Orthodontic appliance and group standardized brackets therefor and methods of making, assembling and using appliance to straighten teeth |
US5395238A (en) | 1990-01-19 | 1995-03-07 | Ormco Corporation | Method of forming orthodontic brace |
US5447432A (en) | 1990-01-19 | 1995-09-05 | Ormco Corporation | Custom orthodontic archwire forming method and apparatus |
US5474448A (en) | 1990-01-19 | 1995-12-12 | Ormco Corporation | Low profile orthodontic appliance |
US5431562A (en) | 1990-01-19 | 1995-07-11 | Ormco Corporation | Method and apparatus for designing and forming a custom orthodontic appliance and for the straightening of teeth therewith |
US5454717A (en) | 1990-01-19 | 1995-10-03 | Ormco Corporation | Custom orthodontic brackets and bracket forming method and apparatus |
US5139419A (en) | 1990-01-19 | 1992-08-18 | Ormco Corporation | Method of forming an orthodontic brace |
JP2784690B2 (en) | 1990-03-13 | 1998-08-06 | コムデント ゲゼルシヤフト ミツト ベシユレンクテル ハフツング | Mouth space measuring method and apparatus for implementing the method |
US5257184A (en) | 1990-04-10 | 1993-10-26 | Mushabac David R | Method and apparatus with multiple data input stylii for collecting curvilinear contour data |
US5562448A (en) | 1990-04-10 | 1996-10-08 | Mushabac; David R. | Method for facilitating dental diagnosis and treatment |
US5569578A (en) | 1990-04-10 | 1996-10-29 | Mushabac; David R. | Method and apparatus for effecting change in shape of pre-existing object |
US5347454A (en) | 1990-04-10 | 1994-09-13 | Mushabac David R | Method, system and mold assembly for use in preparing a dental restoration |
US5545039A (en) | 1990-04-10 | 1996-08-13 | Mushabac; David R. | Method and apparatus for preparing tooth or modifying dental restoration |
US5224049A (en) | 1990-04-10 | 1993-06-29 | Mushabac David R | Method, system and mold assembly for use in preparing a dental prosthesis |
US5691905A (en) | 1990-06-11 | 1997-11-25 | Dentsply Research & Development Corp. | Prosthetic teeth and mold making and polishing therefor |
US5452219A (en) | 1990-06-11 | 1995-09-19 | Dentsply Research & Development Corp. | Method of making a tooth mold |
US5340309A (en) | 1990-09-06 | 1994-08-23 | Robertson James G | Apparatus and method for recording jaw motion |
ES2080206T3 (en) | 1990-10-10 | 1996-02-01 | Mikrona Technologie Ag | RAW PIECE FOR THE MANUFACTURE OF A MACHINED PIECE OF DENTAL TECHNIQUE AND CLAMPING DEVICE FOR THE SAME. |
US5198877A (en) | 1990-10-15 | 1993-03-30 | Pixsys, Inc. | Method and apparatus for three-dimensional non-contact shape sensing |
SE468198B (en) | 1990-12-12 | 1992-11-23 | Nobelpharma Ab | PROCEDURE AND DEVICE FOR MANUFACTURE OF INDIVIDUALLY DESIGNED THREE-DIMENSIONAL BODIES USEFUL AS TENDERS, PROTESTES, ETC |
SE469158B (en) | 1991-11-01 | 1993-05-24 | Nobelpharma Ab | DENTAL SENSOR DEVICE INTENDED TO BE USED IN CONNECTION WITH CONTROL OF A WORKING EQUIPMENT |
CH686657A5 (en) | 1991-11-17 | 1996-05-31 | Liconic Ag | Process for producing a partial replacement of a tooth and means for implementing the method. |
JPH05269146A (en) | 1992-03-23 | 1993-10-19 | Nikon Corp | Extracting method for margin line at the time of designing crown |
US5273429A (en) | 1992-04-03 | 1993-12-28 | Foster-Miller, Inc. | Method and apparatus for modeling a dental prosthesis |
US5378154A (en) | 1992-04-06 | 1995-01-03 | Elephant Holding B.V. | Dental prosthesis and method for manufacturing a dental prosthesis |
NL9200642A (en) | 1992-04-06 | 1993-11-01 | Elephant Holding Bv | METHOD FOR MANUFACTURING A DENTAL PROSTHESIS |
DE4214876C2 (en) | 1992-05-05 | 2000-07-06 | Kaltenbach & Voigt | Optical measurement of teeth without a matt surface treatment |
FR2693096B1 (en) | 1992-07-06 | 1994-09-23 | Deshayes Marie Josephe | Process for modeling the cranial and facial morphology from an x-ray of the skull. |
DE4229466C2 (en) * | 1992-09-03 | 2001-04-26 | Kaltenbach & Voigt | Tooth measurement without calibration body |
AU5598894A (en) | 1992-11-09 | 1994-06-08 | Ormco Corporation | Custom orthodontic appliance forming method and apparatus |
SE501410C2 (en) | 1993-07-12 | 1995-02-06 | Nobelpharma Ab | Method and apparatus in connection with the manufacture of tooth, bridge, etc. |
SE501411C2 (en) | 1993-07-12 | 1995-02-06 | Nobelpharma Ab | Method and apparatus for three-dimensional body useful in the human body |
NL9301308A (en) | 1993-07-26 | 1995-02-16 | Willem Frederick Van Nifterick | Method of securing a dental prosthesis to implants in a patient's jawbone and using means thereof. |
US5382164A (en) | 1993-07-27 | 1995-01-17 | Stern; Sylvan S. | Method for making dental restorations and the dental restoration made thereby |
US5338198A (en) | 1993-11-22 | 1994-08-16 | Dacim Laboratory Inc. | Dental modeling simulator |
SE502035C2 (en) | 1993-12-06 | 1995-07-24 | Nobelpharma Ab | Method and apparatus for producing information for the production of artificial support organs or replacement parts for the human body |
SE502427C2 (en) | 1994-02-18 | 1995-10-16 | Nobelpharma Ab | Method and device utilizing articulator and computer equipment |
JPH07234113A (en) * | 1994-02-24 | 1995-09-05 | Mazda Motor Corp | Method and apparatus for detecting shape |
SE503498C2 (en) | 1994-10-04 | 1996-06-24 | Nobelpharma Ab | Method and device for a product intended to be part of the human body and a scanning device for a model for the product |
US5549476A (en) | 1995-03-27 | 1996-08-27 | Stern; Sylvan S. | Method for making dental restorations and the dental restoration made thereby |
EP0840574B1 (en) | 1995-07-21 | 2003-02-19 | Cadent Ltd. | Method and system for acquiring three-dimensional teeth image |
JP2879003B2 (en) | 1995-11-16 | 1999-04-05 | 株式会社生体光情報研究所 | Image measurement device |
ATE274860T1 (en) | 1995-12-19 | 2004-09-15 | Ivoclar Vivadent Ag | METHOD FOR PRODUCING DENTAL CROWNS AND/OR DENTAL BRIDGES |
US5725376A (en) | 1996-02-27 | 1998-03-10 | Poirier; Michel | Methods for manufacturing a dental implant drill guide and a dental implant superstructure |
US6382975B1 (en) * | 1997-02-26 | 2002-05-07 | Technique D'usinage Sinlab Inc. | Manufacturing a dental implant drill guide and a dental implant superstructure |
US6044170A (en) | 1996-03-21 | 2000-03-28 | Real-Time Geometry Corporation | System and method for rapid shape digitizing and adaptive mesh generation |
US5831719A (en) | 1996-04-12 | 1998-11-03 | Holometrics, Inc. | Laser scanning system |
DE19618140A1 (en) * | 1996-05-06 | 1997-11-13 | Fraunhofer Ges Forschung | 3D measuring arrangement for whole body detection and measurement of a corresponding measuring arrangement |
US5823778A (en) | 1996-06-14 | 1998-10-20 | The United States Of America As Represented By The Secretary Of The Air Force | Imaging method for fabricating dental devices |
US5870220A (en) | 1996-07-12 | 1999-02-09 | Real-Time Geometry Corporation | Portable 3-D scanning system and method for rapid shape digitizing and adaptive mesh generation |
US5812269A (en) | 1996-07-29 | 1998-09-22 | General Scanning, Inc. | Triangulation-based 3-D imaging and processing method and system |
JPH1075963A (en) | 1996-09-06 | 1998-03-24 | Nikon Corp | Method for designing dental prosthetic appliance model and medium recording program for executing the method |
DE19638758A1 (en) * | 1996-09-13 | 1998-03-19 | Rubbert Ruedger | Method and device for three-dimensional measurement of objects |
AUPO280996A0 (en) | 1996-10-04 | 1996-10-31 | Dentech Investments Pty Ltd | Creation and utilization of 3D teeth models |
AU1198797A (en) * | 1996-12-20 | 1998-07-17 | Elypse | Method for producing a dental prosthesis |
US5813859A (en) | 1997-01-23 | 1998-09-29 | Hajjar; Victor J. | Method and apparatus for tooth restoration |
US6217334B1 (en) | 1997-01-28 | 2001-04-17 | Iris Development Corporation | Dental scanning method and apparatus |
IL120135A0 (en) * | 1997-02-03 | 1997-06-10 | Dentop Systems Ltd | A video system for three dimensional imaging and photogrammetry |
US6175415B1 (en) | 1997-02-19 | 2001-01-16 | United Technologies Corporation | Optical profile sensor |
SE509005C2 (en) * | 1997-02-24 | 1998-11-23 | Dentronic Ab | Method and arrangement for non-contact measurement of the three-dimensional shape of detail objects |
DE29705934U1 (en) | 1997-04-03 | 1997-06-05 | Kaltenbach & Voigt | Diagnostic and treatment device for teeth |
US5879158A (en) | 1997-05-20 | 1999-03-09 | Doyle; Walter A. | Orthodontic bracketing system and method therefor |
US5975893A (en) | 1997-06-20 | 1999-11-02 | Align Technology, Inc. | Method and system for incrementally moving teeth |
US6152731A (en) | 1997-09-22 | 2000-11-28 | 3M Innovative Properties Company | Methods for use in dental articulation |
US5882192A (en) | 1997-10-30 | 1999-03-16 | Ortho-Tain, Inc. | Computerized orthodontic diagnosis and appliance dispenser |
DK0913130T3 (en) | 1997-10-31 | 2003-06-16 | Dcs Forschungs & Entwicklungs | Method and apparatus for making a tooth replacement part |
DE19838238A1 (en) * | 1998-08-22 | 2000-03-02 | Girrbach Dental Gmbh | Process for the computer-controlled manufacture of dentures |
US6227850B1 (en) | 1999-05-13 | 2001-05-08 | Align Technology, Inc. | Teeth viewing system |
US6406292B1 (en) * | 1999-05-13 | 2002-06-18 | Align Technology, Inc. | System for determining final position of teeth |
AU2164100A (en) * | 1998-12-04 | 2000-06-26 | Align Technology, Inc. | Reconfigurable dental model system for fabrication of dental appliances |
US6594539B1 (en) * | 1999-03-29 | 2003-07-15 | Genex Technologies, Inc. | Three-dimensional dental imaging method and apparatus having a reflective member |
US6102696A (en) * | 1999-04-30 | 2000-08-15 | Osterwalder; J. Martin | Apparatus for curing resin in dentistry |
JP2000329519A (en) * | 1999-05-18 | 2000-11-30 | Nkk Corp | Coil position detector |
EP1067362A1 (en) * | 1999-07-09 | 2001-01-10 | Hewlett-Packard Company | Document imaging system |
US6350120B1 (en) * | 1999-11-30 | 2002-02-26 | Orametrix, Inc. | Method and apparatus for designing an orthodontic apparatus to provide tooth movement |
US6250918B1 (en) | 1999-11-30 | 2001-06-26 | Orametrix, Inc. | Method and apparatus for simulating tooth movement for an orthodontic patient |
US6402707B1 (en) * | 2000-06-28 | 2002-06-11 | Denupp Corporation Bvi | Method and system for real time intra-orally acquiring and registering three-dimensional measurements and images of intra-oral objects and features |
US6386878B1 (en) * | 2000-08-16 | 2002-05-14 | Align Technology, Inc. | Systems and methods for removing gingiva from teeth |
US6364660B1 (en) * | 2000-10-25 | 2002-04-02 | Duane Milford Durbin | Method and system for imaging and modeling dental structures |
US6386867B1 (en) * | 2000-11-30 | 2002-05-14 | Duane Milford Durbin | Method and system for imaging and modeling dental structures |
JP4012710B2 (en) * | 2001-02-14 | 2007-11-21 | 株式会社リコー | Image input device |
JP2003050112A (en) * | 2001-08-07 | 2003-02-21 | Minolta Co Ltd | Three-dimensional shape input device and projector |
JP2003057021A (en) * | 2001-08-16 | 2003-02-26 | Minolta Co Ltd | Three-dimensional shape input apparatus and projection apparatus |
US20030045798A1 (en) * | 2001-09-04 | 2003-03-06 | Richard Hular | Multisensor probe for tissue identification |
DE60326881D1 (en) | 2002-12-31 | 2009-05-07 | D4D Technologies Llc | DIGITIZATION SYSTEM WITH A LASER FOR DENTAL APPLICATIONS |
-
2004
- 2004-09-17 AU AU2004273957A patent/AU2004273957B2/en active Active
- 2004-09-17 JP JP2006527048A patent/JP4913597B2/en active Active
- 2004-09-17 EP EP04784343A patent/EP1663046A4/en not_active Withdrawn
- 2004-09-17 US US10/943,273 patent/US7342668B2/en active Active
- 2004-09-17 CA CA002538162A patent/CA2538162C/en active Active
- 2004-09-17 WO PCT/US2004/030453 patent/WO2005027770A2/en active Application Filing
Also Published As
Publication number | Publication date |
---|---|
WO2005027770A2 (en) | 2005-03-31 |
AU2004273957B2 (en) | 2009-05-28 |
US20050099638A1 (en) | 2005-05-12 |
EP1663046A4 (en) | 2011-10-05 |
JP4913597B2 (en) | 2012-04-11 |
AU2004273957A1 (en) | 2005-03-31 |
EP1663046A2 (en) | 2006-06-07 |
JP2007521491A (en) | 2007-08-02 |
US7342668B2 (en) | 2008-03-11 |
CA2538162A1 (en) | 2005-03-31 |
WO2005027770A3 (en) | 2008-04-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA2538162C (en) | High speed multiple line three-dimensional digitization | |
KR101601331B1 (en) | System and method for three-dimensional measurment of the shape of material object | |
Gühring | Dense 3D surface acquisition by structured light using off-the-shelf components | |
JP3417377B2 (en) | Three-dimensional shape measuring method and apparatus, and recording medium | |
Zhang et al. | Rapid shape acquisition using color structured light and multi-pass dynamic programming | |
US5175601A (en) | High-speed 3-D surface measurement surface inspection and reverse-CAD system | |
US6611617B1 (en) | Scanning apparatus and method | |
US7061628B2 (en) | Non-contact apparatus and method for measuring surface profile | |
US20060072123A1 (en) | Methods and apparatus for making images including depth information | |
US6611344B1 (en) | Apparatus and method to measure three dimensional data | |
US20150054946A1 (en) | Real-time inspection guidance of triangulation scanner | |
US20090080036A1 (en) | Scanner system and method for scanning | |
EP1643210A1 (en) | Method and apparatus for measuring shape of an object | |
US5090811A (en) | Optical radius gauge | |
EP1680689B1 (en) | Device for scanning three-dimensional objects | |
US6219063B1 (en) | 3D rendering | |
Sansoni et al. | OPL-3D: A novel, portable optical digitizer for fast acquisition of free-form surfaces | |
JPH1114327A (en) | Three-dimensional shape measuring method and device therefor | |
Schoenfeld et al. | Reverse engineering using optical 3D sensors | |
KR100379948B1 (en) | Three-Dimensional Shape Measuring Method | |
Ricci et al. | High-resolution laser radar for 3D imaging in artwork cataloging, reproduction, and restoration | |
JP2002031511A (en) | Three-dimensional digitizer | |
CA2013337C (en) | Optical radius gauge | |
Wetzler et al. | Low cost 3D Laser Scanning Unit with application to Face Recognition | |
McDonald et al. | 3D measurement using stereo scene coding |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request |