US20070206242A1 - Method, apparatus and system providing an integrated hyperspectral imager - Google Patents
Method, apparatus and system providing an integrated hyperspectral imager Download PDFInfo
- Publication number
- US20070206242A1 US20070206242A1 US11/642,867 US64286706A US2007206242A1 US 20070206242 A1 US20070206242 A1 US 20070206242A1 US 64286706 A US64286706 A US 64286706A US 2007206242 A1 US2007206242 A1 US 2007206242A1
- Authority
- US
- United States
- Prior art keywords
- hyperspectral
- imager apparatus
- bandpass
- pixel arrays
- imager
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title abstract description 3
- 238000003491 array Methods 0.000 claims abstract description 62
- 238000003384 imaging method Methods 0.000 claims description 41
- 230000003595 spectral effect Effects 0.000 claims description 11
- 238000001228 spectrum Methods 0.000 claims description 9
- 230000003287 optical effect Effects 0.000 claims description 3
- 238000009501 film coating Methods 0.000 claims 6
- 239000010409 thin film Substances 0.000 claims 6
- 230000004888 barrier function Effects 0.000 claims 2
- 239000000758 substrate Substances 0.000 abstract description 7
- 239000002131 composite material Substances 0.000 abstract 1
- 238000006243 chemical reaction Methods 0.000 description 7
- 239000004065 semiconductor Substances 0.000 description 4
- 238000000701 chemical imaging Methods 0.000 description 3
- 239000010410 layer Substances 0.000 description 3
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000005670 electromagnetic radiation Effects 0.000 description 2
- 239000011229 interlayer Substances 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 229910052710 silicon Inorganic materials 0.000 description 2
- 239000010703 silicon Substances 0.000 description 2
- 230000004913 activation Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 239000011248 coating agent Substances 0.000 description 1
- 238000000576 coating method Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 239000003989 dielectric material Substances 0.000 description 1
- -1 e.g. Substances 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005530 etching Methods 0.000 description 1
- 238000002347 injection Methods 0.000 description 1
- 239000007924 injection Substances 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 238000000465 moulding Methods 0.000 description 1
- 238000012634 optical imaging Methods 0.000 description 1
- 238000002161 passivation Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000000985 reflectance spectrum Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000006641 stabilisation Effects 0.000 description 1
- 238000011105 stabilization Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1462—Coatings
- H01L27/14621—Colour filter arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
- H04N23/13—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths with multiple sensors
- H04N23/15—Image signal generation with circuitry for avoiding or correcting image misregistration
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1462—Coatings
- H01L27/14623—Optical shielding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/40—Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
- H04N25/41—Extracting pixel data from a plurality of image sensors simultaneously picking up an image, e.g. for increasing the field of view by combining the outputs of a plurality of sensors
Definitions
- Embodiments of the invention relate generally to the field of semiconductor devices and more particularly to multi-array image sensors.
- Image sensors such as multispectral image sensors, generally produce images with a few relatively broad wavelength bands from a wavelength of about 400 nm to about 700 nm. These bands typically correlate to the red, green and blue color filters (RGB) of a Bayer patterned color filter array (described below) used in the image sensor.
- RGB red, green and blue color filters
- Hyperspectral image sensors simultaneously collect image data in dozens or hundreds of narrow, adjacent hyperspectral bands. Hyperspectral sensors create a larger number of images from contiguous, rather than disjoint, regions of the spectrum, typically, with much finer resolution than can be obtained with a multispectral image sensor. Hyperspectral imaging involves acquisition of image data in many contiguous narrow hyperspectral bands, the goal being to produce laboratory quality reflectance spectra for each pixel in an image.
- CCDs charge coupled devices
- APS CMOS active pixel sensors
- photodiode arrays charge injection devices
- hybrid focal plane arrays among others.
- CCDs charge coupled devices
- photodiode arrays charge injection devices
- hybrid focal plane arrays among others.
- image sensors use imaging lenses to focus electromagnetic radiation onto photo-conversion devices, e.g., photodiodes.
- these image sensors can use color filters to pass particular wavelengths of electromagnetic radiation to the photo-conversion devices, such that the photo-conversion devices typically are associated with a particular color.
- FIGS. 1A and 1B respectively show a top view and a simplified cross sectional view of a portion of a conventional color image sensor for the visual light spectrum using a Bayer patterned color filter pixel cell array 100 .
- the array 100 includes pixel cells 10 , each being formed on a substrate 1 .
- the pixel array 100 is covered by a color filter array 30 .
- the color filter array 30 includes color filters 31 r, 31 g, 31 b, each disposed over a respective pixel cell 10 .
- Each of the filters 31 r, 31 g, 31 b allows only particular wavelengths of light to pass through to a respective photo-conversion device.
- each pixel cell 10 includes a photo-conversion device 12 r, 12 g, for example, a photodiode, having an associated charge collecting well 13 r, 13 g.
- the elements 12 r, 12 g, 13 r, 13 g are associated with red and green colors based on the color being passed to the pixels of one row of the color filter array 30 , however, it should be appreciated that there may be a photo-conversion device 12 and a charge collecting well 13 that is associated with the color blue that is not shown in the cross sectional view of FIG. 1B .
- the illustrated array 100 has imaging lenses 20 that collect and focus light onto respective photo-conversion devices 12 r, 12 g, which in turn convert the focused light into electrons that are stored in the respective charge collecting wells 13 r, 13 g.
- a passivation layer 6 which typically covers the gate structure of transistors of the pixels and an overlying interlayer dielectric (ILD) region 3 .
- the ILD region 3 typically includes multiple layers of interlayer dielectrics and conductors that form connections between devices of the pixel cells 10 and from the pixel cells 10 to circuitry 150 peripheral to the array 100 .
- a dielectric layer 5 may also be provided between the color filter array 30 and imaging lenses 20 .
- a hyperspectral image sensor or camera system relies on many narrow hyperspectral bandpass filters to capture the hyperspectral image content of a scene.
- the hyperspectral bandpass filters may be applied to an imaging system of the type illustrated in FIG. 1B .
- One major disadvantage of this is that because of the large number of narrow band filters required, the hyperspectral image sensor becomes very large and expensive. Further, because the image sensor relies on applying separate hyperspectral bandpass filters to a lens system over one or more pixels in each pixel array, the filters do not capture the full image content and may produce poor quality images.
- FIG. 1A is a top plan view of a portion of a conventional Bayer pattern color image sensor.
- FIG. 1B is a cross sectional view of a row portion of a conventional Bayer pattern color image sensor.
- FIG. 2A is a plan view of a multi-array image sensor in accordance with an embodiment described herein.
- FIG. 2B is a cross sectional view along a line A-A of the FIG. 2A image sensor in accordance with an embodiment described herein.
- FIG. 3 is a graph showing the measurements of hyperspectral imaging in accordance with an embodiment described herein.
- FIG. 4 is a graph showing the measurements of a continuous spectrum for each hyperspectral image from each pixel in accordance with an embodiment described herein.
- FIG. 5 is a cross sectional view of a portion of an image sensor in accordance to an embodiment described herein.
- FIG. 6 illustrates a block diagram of a CMOS image sensor constructed in accordance with an embodiment described herein.
- FIG. 7 depicts a processor system constructed in accordance with an embodiment described herein.
- the embodiments described herein relate to methods, apparatuses, and systems for integrating a plurality of pixel arrays onto a single substrate, each having an associated hyperspectral bandpass filter and imaging lens.
- Each imaging array and its associated lens and bandpass filter form an image of objects in a scene. If the image sensor arrays and lenses are in close proximity, on the order of less than 2 mm, e.g., about 0.5 mm to about 1 mm, then object parallax will be quite small and each array shared will image the same scene for objects that are all farther than 1 m from the arrays.
- a hyperspectral image may be constructed from individual images respectively acquired by the pixel arrays such that each pixel in the image contains a hyperspectral representation of reflectance of that point in the scene.
- FIG. 2A is a plan view of an embodiment of a hyperspectral image sensor 200 on a semiconductor, e.g., silicon, substrate 201 of a single die.
- image sensor 200 is formed as a 4 ⁇ 4 array configuration of pixel arrays 203 for a total of 16 pixel arrays 203 , each having an associated lens element 204 .
- the imaging lenses 204 are formed by molding or by etching a lens material layer.
- a hyperspectral bandpass filter 205 ( FIG. 2B ) may be formed over each optical imaging lens 204 .
- the hyperspectral bandpass filters 205 may be formed separate from lenses 204 , as illustrated in FIG.
- Each hyperspectral bandpass filter 205 associated with an imaging lens 204 has a different narrow hyperspectral bandpass characteristic.
- there are sixteen different narrow hyperspectral bandpass filters 205 each passing a narrow wavelength range of light to a respective pixel array 203 below.
- Hyperspectral bandpass filters 205 may pass narrow bandwidths that are narrower than the bandwidths used to simply separate red, green and blue wavelengths and may have pass bands of and provide a spectral resolution of less than 100 nm per band, e.g., 50-100 nm per band, and could include infrared bands from about 700 nm to about 1000 nm and/or UV bands from about 200 nm to about 400 nm.
- the filters 205 may also have a bandwidth resolution that is as narrow as less than 10 nm per band, for example, 5-10 nm per band.
- image sensor 200 includes a substrate 201 on which a plurality of pixel arrays 203 and associated support circuitry (not shown) for each array 203 are fabricated.
- the illustrated portion shows one pixel array 203 for one color that has a corresponding color filter that only allows a limited narrow spectrum of wavelengths to pass through.
- Each imaging lens 204 or pixel array 203 then captures different narrow band hyperspectral information.
- the embodiment shown in FIG. 2B shows pixel arrays 203 that are four pixels wide, but it should be appreciated that each of the pixel arrays 203 may contain hundreds, thousands or millions of pixels, as desired.
- Opaque walls 260 separate the individual arrays 203 and opaque walls 270 separate the arrays of imaging lenses 204 that are arranged above each respective pixel array 203 . Opaque walls may also optionally separate the narrow hyperspectral filters 205 .
- the hyperspectral bandpass filters 205 are unique for each imaging lens 204 , meaning that each imaging lens 204 or pixel array 203 under each lens 204 has a different narrow band color filter that only allows a limited spectrum of wavelengths to pass through. Thus, each pixel array 203 captures different narrow band hyperspectral information.
- the pixel arrays 203 preferably are integrated on a single silicon die substrate with common circuitry. Including a multi-array color image sensor on a single die provides for a reduction of color crosstalk artifacts, especially for compact camera modules with pixel sizes less than 6 microns by 6 microns. Moreover, an imaging lens with a short focal length can minimize parallax effects and allow a camera module employing image sensor 200 to be more compact.
- FIG. 2B illustrates a single lens 204 over each array 203
- a modified embodiment provides an imaging lens 204 over one or more pixels in each array 203 .
- each individual lens 204 may cover and focus light on a sixteen pixel section (in a 4 ⁇ 4 pattern) of the pixels in each array 203 .
- the image sensor 200 employs hyperspectral imaging, the image data is simultaneously collected in dozens or hundreds of narrow, adjacent hyperspectral bands, as illustrated in FIG. 3 .
- a 4 ⁇ 4 arrangement of pixel arrays (as illustrated in FIGS. 2A-2B ) has at most 16 hyperspectral bands.
- the image sensor must have a hundred or more individual pixel arrays, e.g., a 10 ⁇ 10 arrangement of pixel arrays.
- Each image I 1 . . . I n from a respective array represents a hyperspectral image from a respective pixel array 203 ( FIG. 2A ).
- the totality of images I 1 . . . I n represent an image with all hyperspectral bands, as illustrated in FIG. 4 .
- the individual images can then be combined and thus, each pixel in the combined image would contain full hyperspectral information.
- ⁇ 1 . . . ⁇ n represents the bandpass optical filter for each image I 1 . . . I n of FIG. 4 .
- Each imaging lens 204 with its unique hyperspectral bandpass filter will image the same scene if the image sensor arrays 203 are in close proximity to one another, e.g., on the order of less than 2 mm and with a preferred spacing in the range of about 0.5 mm to about 1 mm.
- the lenses are similarly spaced.
- a hyperspectral or full spectral image may be constructed such that each pixel in the produced image contains a full spectral representation of reflectance of that point in the scene.
- object parallax is quite small.
- FIG. 5 is a simplified cross sectional view of a portion of an image sensor 200 according to any embodiment of the invention.
- Sensor 200 includes a substrate 201 on which a plurality of pixel arrays 203 and associated support circuitry (not shown) for each array are fabricated.
- Each color pixel array 203 is associated with a different color on a wavelength spectrum.
- each imaging lens 204 has a unique hyperspectral bandpass filter 205 that is associated with each pixel array 203 which is designed to pass a limited spectrum of wavelengths.
- Each pixel array 203 then captures different hyperspectral information.
- N is the pixel array dimension (in pixels)
- d x is the distance between the outer focal points of lenses 204 and where x is approximated by the following formula:
- O is the object distance from the camera (or distance from the object to be imaged and the imaging lenses 204 ) and ⁇ is the field of view (FOV) angle.
- FOV field of view
- the P x or parallax distance at the object plane between adjacent pixel arrays 203 would equal 0.78 pixels by using formulas (1) and (2) above.
- the maximum parallax (at 60° FOV) between images for objects with a distance of greater than or equal to 2 m from the camera would equal 2.35 pixels.
- a 4 ⁇ 4 image sensor would have 16 hyperspectral bands, a pixel element size of 2.2 ⁇ m ⁇ 2.2 ⁇ m and a target hyperspectral image size of 800 ⁇ 600.
- Various image sensor configurations allow tradeoffs in performance. For example, a 3 ⁇ 3 image sensor will have less maximum parallax than a 4 ⁇ 4 image sensor, but will have only 9 hyperspectral bands. A larger FOV will have less parallax but the user must get closer to objects they want to isolate in a scene. Further, smaller pixels may be used to reduce maximum parallax with a tradeoff of a reduction in camera sensitivity.
- FIG. 6 illustrates a block diagram of a CMOS imaging device 500 , which employs a multi-array sensor 200 having a plurality of pixel arrays 203 constructed according to any embodiment described above. Multiple arrays 203 are arranged to form one large pixel array 505 . The pixels of each row in array 505 are all turned on at the same time by a row select line, and the pixels of each column are selectively output onto the output lines by respective column select lines. A plurality of row and column select lines are provided for the entire array 505 . The row lines are selectively activated in sequence by a row driver 510 in response to row address decoder 515 . The column select lines are selectively activated in sequence for each row activation by a column driver 520 in response to column address decoder 525 . Thus, a row and column address is provided for each pixel.
- the CMOS image sensor 500 is operated by a control circuit 530 , which controls address decoders 515 , 525 for selecting the appropriate row and column select lines for pixel readout.
- Control circuit 530 also controls the row and column driver circuitry 510 , 520 so that they apply driving voltages to the drive transistors of the selected row and column select lines.
- the pixel output signals typically include a pixel reset signal V rst read out of the storage region after it is reset by the reset transistor and a pixel image signal V sig , which is read out of the storage region after photo-generated charges are transferred to the region.
- V rst and V sig signals are sampled by a sample and hold circuit 535 and are subtracted by a differential amplifier 540 , to produce a differential signal V rst -V sig for each readout pixel.
- V rst -V sig represents the amount of light impinging on the pixels.
- This difference signal is digitized by an analog-to-digital converter 545 .
- the digitized pixel signals are fed to an image processor 550 to form a digital image output.
- the digitizing and image processing can be located on or off the imaging device chip.
- the differential signal V rst -V sig can be amplified as a differential signal and directly digitized by a differential analog to digital converter.
- the analog-to digital converter 545 supplies the digitized pixel signals to an image processor 550 , which performs appropriate image processing, which can include combining the outputs of multiple arrays and performing a parallax adjustment calculation if needed or desired, before providing digital signals defining an image output.
- FIG. 6 represents one example of a readout circuit for multi-array 505 .
- Another embodiment could employ the readout circuit in FIG. 6 for each individual array 203 with an image processor 550 combining the outputs of the individual arrays 203 .
- FIG. 6 illustrates a readout circuit suitable for CMOS arrays 203 , but the invention may be used with other solid state imaging arrays, for example, CCD arrays in which case a readout circuit suitable for reading out CCD arrays could be employed.
- FIG. 7 illustrates a processor system 900 including the image sensor 500 of FIG. 6 .
- the processor system 900 is one example of a system having digital circuits that could include image sensor devices. Without being limiting, such a processor system could also include a computer system, camera system, scanner, machine vision, vehicle navigation, video phone, surveillance system, auto focus system, star tracker system, motion detection system, image stabilization system, and other processor system.
- the processor system 900 for example a digital camera system, generally comprises a central processing unit (CPU) 995 , such as a microprocessor for common operational control, that communicates with an input/output (I/O) device 991 over a bus 993 .
- Image sensor 500 also communicates with the CPU 995 over the bus 993 .
- the processor-based system 900 also includes random access memory (RAM) 992 , and can include removable memory 994 , such as flash memory, which also communicate with CPU 995 over the bus 993 .
- Image sensor 500 may be combined with a processor, such as a CPU, digital signal processor, or microprocessor, with or without memory storage on a single integrated circuit or on a different chip than the processor.
- a parallax adjustment calculation may be performed by the image processor 550 in image sensor 500 , or by the CPU 995 .
Abstract
Methods, apparatuses, and systems are disclosed which provide a plurality of pixel arrays on a common substrate each associated with a hyperspectral filter. Images from each of the arrays may be separately analyzed or combined into a composite image.
Description
- This application is a continuation-in-part of U.S. application Ser. No. 11/367,580 filed Mar. 6, 2006, the disclosure of which is incorporated by reference herein.
- Embodiments of the invention relate generally to the field of semiconductor devices and more particularly to multi-array image sensors.
- Image sensors, such as multispectral image sensors, generally produce images with a few relatively broad wavelength bands from a wavelength of about 400 nm to about 700 nm. These bands typically correlate to the red, green and blue color filters (RGB) of a Bayer patterned color filter array (described below) used in the image sensor. Hyperspectral image sensors, on the other hand, simultaneously collect image data in dozens or hundreds of narrow, adjacent hyperspectral bands. Hyperspectral sensors create a larger number of images from contiguous, rather than disjoint, regions of the spectrum, typically, with much finer resolution than can be obtained with a multispectral image sensor. Hyperspectral imaging involves acquisition of image data in many contiguous narrow hyperspectral bands, the goal being to produce laboratory quality reflectance spectra for each pixel in an image.
- The semiconductor industry currently produces different types of semiconductor-based image sensors, such as charge coupled devices (CCDs), CMOS active pixel sensors (APS), photodiode arrays, charge injection devices and hybrid focal plane arrays, among others. These image sensors use imaging lenses to focus electromagnetic radiation onto photo-conversion devices, e.g., photodiodes. Also, these image sensors can use color filters to pass particular wavelengths of electromagnetic radiation to the photo-conversion devices, such that the photo-conversion devices typically are associated with a particular color.
-
FIGS. 1A and 1B respectively show a top view and a simplified cross sectional view of a portion of a conventional color image sensor for the visual light spectrum using a Bayer patterned color filterpixel cell array 100. Thearray 100 includespixel cells 10, each being formed on asubstrate 1. Thepixel array 100 is covered by a color filter array 30. The color filter array 30 includescolor filters respective pixel cell 10. Each of thefilters green color filters 31 g for everyred color filter 31 r andblue color filter 31 b, arranged as shown inFIG. 1A . As shown inFIG. 1B , eachpixel cell 10 includes a photo-conversion device elements conversion device 12 and a charge collecting well 13 that is associated with the color blue that is not shown in the cross sectional view ofFIG. 1B . The illustratedarray 100 hasimaging lenses 20 that collect and focus light onto respective photo-conversion devices charge collecting wells - Between the color filter array 30 and the
pixel cells 10 is apassivation layer 6 which typically covers the gate structure of transistors of the pixels and an overlying interlayer dielectric (ILD)region 3. The ILDregion 3 typically includes multiple layers of interlayer dielectrics and conductors that form connections between devices of thepixel cells 10 and from thepixel cells 10 to circuitry 150 peripheral to thearray 100. Adielectric layer 5 may also be provided between the color filter array 30 andimaging lenses 20. - As discussed above, a hyperspectral image sensor or camera system relies on many narrow hyperspectral bandpass filters to capture the hyperspectral image content of a scene. The hyperspectral bandpass filters may be applied to an imaging system of the type illustrated in
FIG. 1B . One major disadvantage of this is that because of the large number of narrow band filters required, the hyperspectral image sensor becomes very large and expensive. Further, because the image sensor relies on applying separate hyperspectral bandpass filters to a lens system over one or more pixels in each pixel array, the filters do not capture the full image content and may produce poor quality images. - Therefore, it would be advantageous to have an integrated hyperspectral image sensor which better captures the hyperspectral image content of a scene, which is also compact.
-
FIG. 1A is a top plan view of a portion of a conventional Bayer pattern color image sensor. -
FIG. 1B is a cross sectional view of a row portion of a conventional Bayer pattern color image sensor. -
FIG. 2A is a plan view of a multi-array image sensor in accordance with an embodiment described herein. -
FIG. 2B is a cross sectional view along a line A-A of theFIG. 2A image sensor in accordance with an embodiment described herein. -
FIG. 3 is a graph showing the measurements of hyperspectral imaging in accordance with an embodiment described herein. -
FIG. 4 is a graph showing the measurements of a continuous spectrum for each hyperspectral image from each pixel in accordance with an embodiment described herein. -
FIG. 5 is a cross sectional view of a portion of an image sensor in accordance to an embodiment described herein. -
FIG. 6 illustrates a block diagram of a CMOS image sensor constructed in accordance with an embodiment described herein. -
FIG. 7 depicts a processor system constructed in accordance with an embodiment described herein. - In the following detailed description, reference is made to the accompanying drawings which form a part hereof, and in which are shown by way of illustration of specific embodiments in which the invention may be practiced. These embodiments are described in sufficient detail to enable those of ordinary skill in the art to make and use them, and it is to be understood that structural, logical or procedural changes may be made to the specific embodiments disclosed herein.
- The embodiments described herein relate to methods, apparatuses, and systems for integrating a plurality of pixel arrays onto a single substrate, each having an associated hyperspectral bandpass filter and imaging lens. Each imaging array and its associated lens and bandpass filter form an image of objects in a scene. If the image sensor arrays and lenses are in close proximity, on the order of less than 2 mm, e.g., about 0.5 mm to about 1 mm, then object parallax will be quite small and each array shared will image the same scene for objects that are all farther than 1 m from the arrays. A hyperspectral image may be constructed from individual images respectively acquired by the pixel arrays such that each pixel in the image contains a hyperspectral representation of reflectance of that point in the scene.
-
FIG. 2A is a plan view of an embodiment of ahyperspectral image sensor 200 on a semiconductor, e.g., silicon,substrate 201 of a single die. In the illustrated embodiment,image sensor 200 is formed as a 4×4 array configuration ofpixel arrays 203 for a total of 16pixel arrays 203, each having an associatedlens element 204. Theimaging lenses 204 are formed by molding or by etching a lens material layer. A hyperspectral bandpass filter 205 (FIG. 2B ) may be formed over eachoptical imaging lens 204. Thehyperspectral bandpass filters 205 may be formed separate fromlenses 204, as illustrated inFIG. 2B , and may be formed above or below thelenses 204, or may be formed as a coating on thelenses 204, or may be formed as part of thelens 204 by using a colored material forlens 204. Eachhyperspectral bandpass filter 205 associated with animaging lens 204 has a different narrow hyperspectral bandpass characteristic. Thus, in the embodiment illustrated inFIGS. 2A-2B , there are sixteen different narrow hyperspectralbandpass filters 205, each passing a narrow wavelength range of light to arespective pixel array 203 below. Hyperspectral bandpass filters 205 may pass narrow bandwidths that are narrower than the bandwidths used to simply separate red, green and blue wavelengths and may have pass bands of and provide a spectral resolution of less than 100 nm per band, e.g., 50-100 nm per band, and could include infrared bands from about 700 nm to about 1000 nm and/or UV bands from about 200 nm to about 400 nm. Thefilters 205 may also have a bandwidth resolution that is as narrow as less than 10 nm per band, for example, 5-10 nm per band. - Referring to
FIG. 2B ,image sensor 200 includes asubstrate 201 on which a plurality ofpixel arrays 203 and associated support circuitry (not shown) for eacharray 203 are fabricated. The illustrated portion shows onepixel array 203 for one color that has a corresponding color filter that only allows a limited narrow spectrum of wavelengths to pass through. Eachimaging lens 204 orpixel array 203 then captures different narrow band hyperspectral information. For simplicity, the embodiment shown inFIG. 2B showspixel arrays 203 that are four pixels wide, but it should be appreciated that each of thepixel arrays 203 may contain hundreds, thousands or millions of pixels, as desired.Opaque walls 260 separate theindividual arrays 203 andopaque walls 270 separate the arrays ofimaging lenses 204 that are arranged above eachrespective pixel array 203. Opaque walls may also optionally separate the narrowhyperspectral filters 205. The hyperspectral bandpass filters 205 are unique for eachimaging lens 204, meaning that eachimaging lens 204 orpixel array 203 under eachlens 204 has a different narrow band color filter that only allows a limited spectrum of wavelengths to pass through. Thus, eachpixel array 203 captures different narrow band hyperspectral information. - As noted, the
pixel arrays 203 preferably are integrated on a single silicon die substrate with common circuitry. Including a multi-array color image sensor on a single die provides for a reduction of color crosstalk artifacts, especially for compact camera modules with pixel sizes less than 6 microns by 6 microns. Moreover, an imaging lens with a short focal length can minimize parallax effects and allow a camera module employingimage sensor 200 to be more compact. - Although
FIG. 2B illustrates asingle lens 204 over eacharray 203, a modified embodiment provides animaging lens 204 over one or more pixels in eacharray 203. For example, eachindividual lens 204 may cover and focus light on a sixteen pixel section (in a 4×4 pattern) of the pixels in eacharray 203. - Because the
image sensor 200 employs hyperspectral imaging, the image data is simultaneously collected in dozens or hundreds of narrow, adjacent hyperspectral bands, as illustrated inFIG. 3 . For example, a 4×4 arrangement of pixel arrays (as illustrated inFIGS. 2A-2B ) has at most 16 hyperspectral bands. For collection of hundreds of hyperspectral bands, the image sensor must have a hundred or more individual pixel arrays, e.g., a 10×10 arrangement of pixel arrays. - Each image I1 . . . In from a respective array represents a hyperspectral image from a respective pixel array 203 (
FIG. 2A ). The totality of images I1 . . . In represent an image with all hyperspectral bands, as illustrated inFIG. 4 . The individual images can then be combined and thus, each pixel in the combined image would contain full hyperspectral information. InFIG. 3 , β1 . . . βn represents the bandpass optical filter for each image I1 . . . In ofFIG. 4 . Eachimaging lens 204 with its unique hyperspectral bandpass filter will image the same scene if theimage sensor arrays 203 are in close proximity to one another, e.g., on the order of less than 2 mm and with a preferred spacing in the range of about 0.5 mm to about 1 mm. Forarrays 203 which haverespective lenses 204, the lenses are similarly spaced. Thus, a hyperspectral or full spectral image may be constructed such that each pixel in the produced image contains a full spectral representation of reflectance of that point in the scene. Further, because theimage sensor arrays 203 andimaging lenses 204 are in close proximity, e.g., on the order of less than 2 mm and preferably about 0.5 mm to about 1 mm, then object parallax is quite small. -
FIG. 5 is a simplified cross sectional view of a portion of animage sensor 200 according to any embodiment of the invention.Sensor 200 includes asubstrate 201 on which a plurality ofpixel arrays 203 and associated support circuitry (not shown) for each array are fabricated. Eachcolor pixel array 203 is associated with a different color on a wavelength spectrum. As discussed above, eachimaging lens 204 has a unique hyperspectralbandpass filter 205 that is associated with eachpixel array 203 which is designed to pass a limited spectrum of wavelengths. Eachpixel array 203 then captures different hyperspectral information. - As an object to be imaged moves closer to the array of
imaging lenses 204, theindividual arrays 203 will exhibit an increase in parallax distance between them. The magnitude of the parallax distance between two adjacent arrays is approximated by the following formula: -
P x=(N/x)*d x (1) - where N is the pixel array dimension (in pixels), dx is the distance between the outer focal points of
lenses 204 and where x is approximated by the following formula: -
x=2*O*Tan(α/2) (2) - where O is the object distance from the camera (or distance from the object to be imaged and the imaging lenses 204) and α is the field of view (FOV) angle. Once the object distance O has been measured or approximated, the parallax distance calculation can be performed. An example for calculating parallax is discussed below. This example also shows that small pixel arrays (on the order of about 1 mm and spaced less than 2 mm apart, e.g., from about 0.5 mm to about 1 mm) will not produce excessive parallax.
- For an object distance O of about 2 m from a camera and a field of view angle of 60° , the Px or parallax distance at the object plane between
adjacent pixel arrays 203 would equal 0.78 pixels by using formulas (1) and (2) above. Thus, for image sensors with 4×4 imaging arrays (i.e., 4×4 image sensor), the maximum parallax (at 60° FOV) between images for objects with a distance of greater than or equal to 2 m from the camera would equal 2.35 pixels. A 4×4 image sensor would have 16 hyperspectral bands, a pixel element size of 2.2 μm×2.2 μm and a target hyperspectral image size of 800×600. - Various image sensor configurations allow tradeoffs in performance. For example, a 3×3 image sensor will have less maximum parallax than a 4×4 image sensor, but will have only 9 hyperspectral bands. A larger FOV will have less parallax but the user must get closer to objects they want to isolate in a scene. Further, smaller pixels may be used to reduce maximum parallax with a tradeoff of a reduction in camera sensitivity.
-
FIG. 6 illustrates a block diagram of aCMOS imaging device 500, which employs amulti-array sensor 200 having a plurality ofpixel arrays 203 constructed according to any embodiment described above.Multiple arrays 203 are arranged to form onelarge pixel array 505. The pixels of each row inarray 505 are all turned on at the same time by a row select line, and the pixels of each column are selectively output onto the output lines by respective column select lines. A plurality of row and column select lines are provided for theentire array 505. The row lines are selectively activated in sequence by arow driver 510 in response torow address decoder 515. The column select lines are selectively activated in sequence for each row activation by acolumn driver 520 in response tocolumn address decoder 525. Thus, a row and column address is provided for each pixel. - The
CMOS image sensor 500 is operated by acontrol circuit 530, which controlsaddress decoders Control circuit 530 also controls the row andcolumn driver circuitry circuit 535 and are subtracted by adifferential amplifier 540, to produce a differential signal Vrst-Vsig for each readout pixel. Vrst-Vsig represents the amount of light impinging on the pixels. This difference signal is digitized by an analog-to-digital converter 545. The digitized pixel signals are fed to animage processor 550 to form a digital image output. The digitizing and image processing can be located on or off the imaging device chip. In some arrangements the differential signal Vrst-Vsig can be amplified as a differential signal and directly digitized by a differential analog to digital converter. The analog-todigital converter 545 supplies the digitized pixel signals to animage processor 550, which performs appropriate image processing, which can include combining the outputs of multiple arrays and performing a parallax adjustment calculation if needed or desired, before providing digital signals defining an image output. - It should be noted that
FIG. 6 represents one example of a readout circuit formulti-array 505. Another embodiment could employ the readout circuit inFIG. 6 for eachindividual array 203 with animage processor 550 combining the outputs of theindividual arrays 203. Also,FIG. 6 illustrates a readout circuit suitable forCMOS arrays 203, but the invention may be used with other solid state imaging arrays, for example, CCD arrays in which case a readout circuit suitable for reading out CCD arrays could be employed. -
FIG. 7 illustrates aprocessor system 900 including theimage sensor 500 ofFIG. 6 . Theprocessor system 900 is one example of a system having digital circuits that could include image sensor devices. Without being limiting, such a processor system could also include a computer system, camera system, scanner, machine vision, vehicle navigation, video phone, surveillance system, auto focus system, star tracker system, motion detection system, image stabilization system, and other processor system. - The
processor system 900, for example a digital camera system, generally comprises a central processing unit (CPU) 995, such as a microprocessor for common operational control, that communicates with an input/output (I/O)device 991 over abus 993.Image sensor 500 also communicates with theCPU 995 over thebus 993. The processor-basedsystem 900 also includes random access memory (RAM) 992, and can includeremovable memory 994, such as flash memory, which also communicate withCPU 995 over thebus 993.Image sensor 500 may be combined with a processor, such as a CPU, digital signal processor, or microprocessor, with or without memory storage on a single integrated circuit or on a different chip than the processor. A parallax adjustment calculation may be performed by theimage processor 550 inimage sensor 500, or by theCPU 995. - While the embodiments have been described in detail in connection with the embodiments known at the time, it should be readily understood that the claimed invention is not limited to the disclosed embodiments. Rather, they can be modified to incorporate any number of variations, alterations, substitutions or equivalent arrangements not heretofore described. For example, while embodiments are described in connection with a CMOS pixel image sensor, they can be practiced with any other type of pixel image sensor (e.g., CCD, etc.). Furthermore, the various embodiments could be used in automotive applications and other applications where the object plane is at a constant distance, parallax is easily accounted for by using a simple linear shift of pixel data from each camera to properly register all images, such as in machine vision or industrial imaging. In addition, various embodiments may be used in low-cost, solid-state hyperspectral scanners, in which multiple hyperspectral cameras in one scanning system may produce high-resolution images very quickly.
Claims (48)
1. An imager apparatus comprising:
a plurality of pixel arrays on a single die; and
a plurality of imaging lenses for focusing an image on the pixel arrays, each of the pixel arrays being associated with a respective hyperspectral bandpass filter.
2. The imager apparatus of claim 1 , wherein each of the hyperspectral filters has a spectral bandpass of less than 100 nm.
3. The imager apparatus of claim 2 , wherein each of the hyperspectral filters has a spectral bandpass of between 50 nm and 100 nm.
4. The imager apparatus of claim 2 , wherein each of the hyperspectral filters has a spectral bandpass of less than 10 nm.
5. The imager apparatus of claim 4 , wherein each of the hyperspectral filters has a spectral bandpass of between 5 nm and 10 nm.
6. The imager apparatus of claim 1 , wherein each hyperspectral bandpass filter is integrated with its associated imaging lens.
7. The imager apparatus of claim 6 , wherein each hyperspectral bandpass filter is a thin film coating on its associated lens.
8. The imager apparatus of claim 1 , wherein each hyperspectral bandpass filter is a separate element from its associated imaging lens.
9. The imager apparatus of claim 1 , wherein each hyperspectral bandpass filter is a thin film coating on a surface of its associated pixel array.
10. The imager apparatus of claim 1 , wherein each hyperspectral bandpass filter has a unique bandpass range.
11. The imager apparatus of claim 1 , wherein the hyperspectral bandpass filters collectively cover a spectrum of between about 700 nm and about 1000 nm.
12. The imager apparatus of claim 1 , wherein the hyperspectral bandpass filters collectively cover a spectrum of between about 200 nm and about 400 nm.
13. The imager apparatus of claim 1 , wherein the plurality of pixel arrays comprise M×N pixel arrays arranged in a M×N pattern of arrays on the die.
14. The imager apparatus of claim 13 , wherein M equals N.
15. The imager apparatus of claim 13 , wherein M does not equal N.
16. The imager apparatus of claim 1 , wherein the pixel arrays are spaced less than 2 mm apart.
17. The imager apparatus of claim 16 , wherein the pixel arrays are spaced about 0.5 mm to about 1 mm apart.
18. The imager apparatus of claim 16 , wherein the imaging lenses and associated hyperspectral bandpass filters are configured to capture a full image of a scene.
19. The imager apparatus of claim 1 , further comprising an optical barrier between adjacent pixel arrays.
20. The imager apparatus of claim 1 , further comprising an optical barrier between adjacent imaging lenses.
21. An imager apparatus comprising:
a plurality of pixel arrays formed on a single die, wherein the plurality of arrays are spaced apart by less than 2 millimeters;
a plurality of hyperspectral bandpass filters respectively associated with the pixel arrays, each of the hyperspectral filters having a spectral bandpass of less than 100 nm; and
a plurality of imaging lenses respectively associated with the pixel arrays.
22. The imager apparatus of claim 21 , wherein each of the hyperspectral filters has a spectral bandpass of between 50 nm and 100 nm.
23. The imager apparatus of claim 21 , wherein each of the hyperspectral filters has a spectral bandpass of less than 10 nm.
24. The imager apparatus of claim 23 , wherein each of the hyperspectral filters has a spectral bandpass of between 5 nm and 10 nm.
25. The imager apparatus of claim 21 , wherein each imaging lens has a hyperspectral bandpass filter associated with its respective imaging lens.
26. The imager apparatus of claim 25 , wherein each hyperspectral bandpass filter is integrated with its associated imaging lens.
27. The imager apparatus of claim 26 , wherein each hyperspectral bandpass filter is a thin film coating on its associated lens.
28. The imager apparatus of claim 25 , wherein each hyperspectral bandpass filter is a separate element from its associated imaging lens.
29. The imager apparatus of claim 21 , wherein each hyperspectral bandpass filter is a thin film coating on a surface of its associated pixel array.
30. The imager apparatus of claim 21 , wherein each hyperspectral bandpass filter has a unique bandpass range.
31. The imager apparatus of claim 21 , wherein the plurality of pixel arrays comprise M×N pixel arrays arranged in a M×N pattern of arrays on the die.
32. The imager apparatus of claim 31 , wherein M equals N.
33. The imager apparatus of claim 31 , wherein M does not equal N.
34. The imager apparatus of claim 21 , wherein the pixel arrays and the imaging lenses are spaced about 0.5 mm to about 1 mm apart.
35. The imager apparatus of claim 34 , wherein the imaging lenses and associated hyperspectral bandpass filters are configured to capture a full image of a scene.
36. An imager apparatus comprising:
an image sensor comprising:
a plurality of pixel arrays;
a plurality of imaging lenses respectively arranged above the pixel arrays; and
a hyperspectral bandpass filter associated with each imaging lens, wherein each hyperspectral bandpass filter is unique for each imaging lens; and
a readout circuit for reading out respective image signals from each of the arrays.
37. The imager apparatus of claim 36 , further comprising:
an image combining device for combining the respective image signals from each of the pixel arrays.
38. The imager apparatus of claim 37 , wherein the image combining device performs a parallax adjustment calculation.
39. The imager apparatus of claim 37 , wherein a common readout circuit is provided for all of the pixel arrays.
40. The imager apparatus of claim 37 , wherein separate readout circuits are provided for each of the pixel arrays.
41. The imager apparatus of claim 36 , wherein the plurality of pixel arrays are formed on a single die.
42. The imager apparatus of claim 36 , wherein each hyperspectral bandpass filter is integrated with its associated imaging lens.
43. The imager apparatus of claim 42 , wherein each hyperspectral bandpass filter is a thin film coating on its associated lens.
44. The imager apparatus of claim 36 , wherein each hyperspectral bandpass filter is a separate element from its associated imaging lens.
45. The imager apparatus of claim 36 , wherein each hyperspectral bandpass filter is a thin film coating on a surface of its associated pixel array.
46. The imager apparatus of claim 36 , wherein the pixel arrays are spaced less than 2 mm apart.
47. The imager apparatus of claim 46 , wherein the pixel arrays are spaced about 0.5 mm to about 1 mm apart.
48. The imager apparatus of claim 36 , wherein the imager apparatus is a component of a digital camera.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/642,867 US20070206242A1 (en) | 2006-03-06 | 2006-12-21 | Method, apparatus and system providing an integrated hyperspectral imager |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/367,580 US7924483B2 (en) | 2006-03-06 | 2006-03-06 | Fused multi-array color image sensor |
US11/642,867 US20070206242A1 (en) | 2006-03-06 | 2006-12-21 | Method, apparatus and system providing an integrated hyperspectral imager |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/367,580 Continuation-In-Part US7924483B2 (en) | 2006-03-06 | 2006-03-06 | Fused multi-array color image sensor |
Publications (1)
Publication Number | Publication Date |
---|---|
US20070206242A1 true US20070206242A1 (en) | 2007-09-06 |
Family
ID=38193548
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/367,580 Active 2030-02-10 US7924483B2 (en) | 2006-03-06 | 2006-03-06 | Fused multi-array color image sensor |
US11/642,867 Abandoned US20070206242A1 (en) | 2006-03-06 | 2006-12-21 | Method, apparatus and system providing an integrated hyperspectral imager |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/367,580 Active 2030-02-10 US7924483B2 (en) | 2006-03-06 | 2006-03-06 | Fused multi-array color image sensor |
Country Status (7)
Country | Link |
---|---|
US (2) | US7924483B2 (en) |
EP (1) | EP2008470A1 (en) |
JP (1) | JP2009529291A (en) |
KR (1) | KR20080109794A (en) |
CN (1) | CN101395926A (en) |
TW (1) | TWI398948B (en) |
WO (1) | WO2007103212A1 (en) |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070201114A1 (en) * | 2006-02-28 | 2007-08-30 | Yoshitaka Egawa | Solid-state image sensing device having photoelectric conversion cells each configured by n pixels in vertical direction |
WO2009085305A1 (en) | 2007-12-27 | 2009-07-09 | Google Inc. | High-resolution, variable depth of field image device |
US20120147233A1 (en) * | 2010-12-09 | 2012-06-14 | Sony Corporation | Imaging device and imaging apparatus |
US20140267878A1 (en) * | 2011-11-04 | 2014-09-18 | Imec | Spectral camera with overlapping segments of image copies interleaved onto sensor array |
US20140347533A1 (en) * | 2013-05-21 | 2014-11-27 | Olympus Corporation | Image processing device and image processing method |
US20150077524A1 (en) * | 2012-03-16 | 2015-03-19 | Nikon Corporation | Image sensor and imaging device |
US20150177429A1 (en) * | 2012-06-05 | 2015-06-25 | Hypermed Imaging, Inc. | Single-sensor hyperspectral imaging device |
US20150325613A1 (en) * | 2013-07-30 | 2015-11-12 | Heptagon Micro Optics Pte. Ltd. | Optoelectronic modules that have shielding to reduce light leakage or stray light, and fabrication methods for such modules |
US9217671B2 (en) | 2009-09-01 | 2015-12-22 | Koninklijke Philips N.V. | High spectral resolution color sensor using non-dispersive elements |
US9772229B2 (en) | 2011-11-04 | 2017-09-26 | Imec | Spectral camera with integrated filters and multiple adjacent image copies projected onto sensor array |
US20190190237A1 (en) * | 2016-09-29 | 2019-06-20 | Intel Corporation | Optical free air bus interconnect |
CN110235252A (en) * | 2016-12-07 | 2019-09-13 | 兴业法兰西红外探测器-索弗拉迪尔公司 | For improving the device of multispectral detection |
US20200329205A1 (en) * | 2020-04-02 | 2020-10-15 | Sichuan Dualix Spectral Lmaging Technology Co., Ltd | Hyperspectral camera based on continuously variable film filter and coating method thereof |
US20220214217A1 (en) * | 2021-01-04 | 2022-07-07 | Argo AI, LLC | Systems and methods for characterizing spectral reflectance of real world objects |
Families Citing this family (124)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7955764B2 (en) * | 2006-04-07 | 2011-06-07 | Micron Technology, Inc. | Methods to make sidewall light shields for color filter array |
US7718533B2 (en) * | 2007-05-08 | 2010-05-18 | Micron Technology, Inc. | Inverted variable resistance memory cell and method of making the same |
US20090051790A1 (en) * | 2007-08-21 | 2009-02-26 | Micron Technology, Inc. | De-parallax methods and apparatuses for lateral sensor arrays |
CN101459853B (en) * | 2007-12-10 | 2011-08-10 | 鸿富锦精密工业(深圳)有限公司 | Color processing apparatus and method |
JP5173536B2 (en) * | 2008-04-02 | 2013-04-03 | シャープ株式会社 | Imaging apparatus and optical axis control method |
EP2283644A4 (en) * | 2008-05-09 | 2011-10-26 | Ecole Polytech | Image sensor having nonlinear response |
CN102037717B (en) | 2008-05-20 | 2013-11-06 | 派力肯成像公司 | Capturing and processing of images using monolithic camera array with hetergeneous imagers |
US11792538B2 (en) | 2008-05-20 | 2023-10-17 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US8866920B2 (en) | 2008-05-20 | 2014-10-21 | Pelican Imaging Corporation | Capturing and processing of images using monolithic camera array with heterogeneous imagers |
JP4952768B2 (en) * | 2009-10-27 | 2012-06-13 | 株式会社ニコン | Imaging apparatus and image analysis computer program |
EP2502115A4 (en) | 2009-11-20 | 2013-11-06 | Pelican Imaging Corp | Capturing and processing of images using monolithic camera array with heterogeneous imagers |
US8319855B2 (en) * | 2010-01-19 | 2012-11-27 | Rambus Inc. | Method, apparatus and system for image acquisition and conversion |
KR101824672B1 (en) | 2010-05-12 | 2018-02-05 | 포토네이션 케이맨 리미티드 | Architectures for imager arrays and array cameras |
WO2012051751A1 (en) * | 2010-10-19 | 2012-04-26 | Ether Precision, Inc. | Optical module comprising monochromatic image sensors, system comprising optical module and method of manufacturing optical module |
US8878950B2 (en) | 2010-12-14 | 2014-11-04 | Pelican Imaging Corporation | Systems and methods for synthesizing high resolution images using super-resolution processes |
US20120274811A1 (en) * | 2011-04-28 | 2012-11-01 | Dmitry Bakin | Imaging devices having arrays of image sensors and precision offset lenses |
JP2014519741A (en) | 2011-05-11 | 2014-08-14 | ペリカン イメージング コーポレイション | System and method for transmitting and receiving array camera image data |
JP2012253727A (en) | 2011-06-07 | 2012-12-20 | Toshiba Corp | Solid-state image sensor and camera module |
US20130265459A1 (en) | 2011-06-28 | 2013-10-10 | Pelican Imaging Corporation | Optical arrangements for use with an array camera |
WO2013003276A1 (en) | 2011-06-28 | 2013-01-03 | Pelican Imaging Corporation | Optical arrangements for use with an array camera |
WO2013043751A1 (en) | 2011-09-19 | 2013-03-28 | Pelican Imaging Corporation | Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures |
KR102002165B1 (en) | 2011-09-28 | 2019-07-25 | 포토내이션 리미티드 | Systems and methods for encoding and decoding light field image files |
EP2817955B1 (en) | 2012-02-21 | 2018-04-11 | FotoNation Cayman Limited | Systems and methods for the manipulation of captured light field image data |
US9210392B2 (en) | 2012-05-01 | 2015-12-08 | Pelican Imaging Coporation | Camera modules patterned with pi filter groups |
KR20150023907A (en) | 2012-06-28 | 2015-03-05 | 펠리칸 이매징 코포레이션 | Systems and methods for detecting defective camera arrays, optic arrays, and sensors |
US20140002674A1 (en) | 2012-06-30 | 2014-01-02 | Pelican Imaging Corporation | Systems and Methods for Manufacturing Camera Modules Using Active Alignment of Lens Stack Arrays and Sensors |
EP2888720B1 (en) | 2012-08-21 | 2021-03-17 | FotoNation Limited | System and method for depth estimation from images captured using array cameras |
WO2014032020A2 (en) | 2012-08-23 | 2014-02-27 | Pelican Imaging Corporation | Feature based high resolution motion estimation from low resolution images captured using an array source |
US9214013B2 (en) | 2012-09-14 | 2015-12-15 | Pelican Imaging Corporation | Systems and methods for correcting user identified artifacts in light field images |
US20140092281A1 (en) | 2012-09-28 | 2014-04-03 | Pelican Imaging Corporation | Generating Images from Light Fields Utilizing Virtual Viewpoints |
WO2014078443A1 (en) | 2012-11-13 | 2014-05-22 | Pelican Imaging Corporation | Systems and methods for array camera focal plane control |
US9363425B2 (en) | 2012-12-06 | 2016-06-07 | Semiconductor Components Industries, Llc | Color filter arrangements for fused array imaging systems |
WO2014130849A1 (en) | 2013-02-21 | 2014-08-28 | Pelican Imaging Corporation | Generating compressed light field representation data |
WO2014133974A1 (en) | 2013-02-24 | 2014-09-04 | Pelican Imaging Corporation | Thin form computational and modular array cameras |
WO2014133481A1 (en) * | 2013-02-26 | 2014-09-04 | Hewlett-Packard Development Company, L.P. | Multiview 3d telepresence |
US9774789B2 (en) | 2013-03-08 | 2017-09-26 | Fotonation Cayman Limited | Systems and methods for high dynamic range imaging using array cameras |
US8866912B2 (en) | 2013-03-10 | 2014-10-21 | Pelican Imaging Corporation | System and methods for calibration of an array camera using a single captured image |
US9521416B1 (en) | 2013-03-11 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for image data compression |
US9106784B2 (en) | 2013-03-13 | 2015-08-11 | Pelican Imaging Corporation | Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing |
US9888194B2 (en) | 2013-03-13 | 2018-02-06 | Fotonation Cayman Limited | Array camera architecture implementing quantum film image sensors |
WO2014165244A1 (en) | 2013-03-13 | 2014-10-09 | Pelican Imaging Corporation | Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies |
WO2014164550A2 (en) | 2013-03-13 | 2014-10-09 | Pelican Imaging Corporation | System and methods for calibration of an array camera |
US9100586B2 (en) | 2013-03-14 | 2015-08-04 | Pelican Imaging Corporation | Systems and methods for photometric normalization in array cameras |
WO2014159779A1 (en) | 2013-03-14 | 2014-10-02 | Pelican Imaging Corporation | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
CN104051484B (en) * | 2013-03-15 | 2017-09-19 | 豪威科技股份有限公司 | Possess the imaging sensor of the pixel with increased optical crosstalk and its use |
US10122993B2 (en) | 2013-03-15 | 2018-11-06 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
US9497429B2 (en) | 2013-03-15 | 2016-11-15 | Pelican Imaging Corporation | Extended color processing on pelican array cameras |
US9215430B2 (en) | 2013-03-15 | 2015-12-15 | Omnivision Technologies, Inc. | Image sensor with pixels having increased optical crosstalk |
EP2973476A4 (en) | 2013-03-15 | 2017-01-18 | Pelican Imaging Corporation | Systems and methods for stereo imaging with camera arrays |
US9445003B1 (en) | 2013-03-15 | 2016-09-13 | Pelican Imaging Corporation | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
WO2014150856A1 (en) | 2013-03-15 | 2014-09-25 | Pelican Imaging Corporation | Array camera implementing quantum dot color filters |
US9633442B2 (en) | 2013-03-15 | 2017-04-25 | Fotonation Cayman Limited | Array cameras including an array camera module augmented with a separate camera |
US20150022655A1 (en) * | 2013-07-19 | 2015-01-22 | Forrest R. Ruhge | Apparatus and method using a linear array of optical sensors for imaging a rotating component of a gas turbine engine |
US9898856B2 (en) | 2013-09-27 | 2018-02-20 | Fotonation Cayman Limited | Systems and methods for depth-assisted perspective distortion correction |
US9264592B2 (en) | 2013-11-07 | 2016-02-16 | Pelican Imaging Corporation | Array camera modules incorporating independently aligned lens stacks |
WO2015074078A1 (en) | 2013-11-18 | 2015-05-21 | Pelican Imaging Corporation | Estimating depth from projected texture using camera arrays |
WO2015081279A1 (en) | 2013-11-26 | 2015-06-04 | Pelican Imaging Corporation | Array camera configurations incorporating multiple constituent array cameras |
WO2015134996A1 (en) | 2014-03-07 | 2015-09-11 | Pelican Imaging Corporation | System and methods for depth regularization and semiautomatic interactive matting using rgb-d images |
US9521319B2 (en) | 2014-06-18 | 2016-12-13 | Pelican Imaging Corporation | Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor |
US10250871B2 (en) | 2014-09-29 | 2019-04-02 | Fotonation Limited | Systems and methods for dynamic calibration of array cameras |
US9942474B2 (en) | 2015-04-17 | 2018-04-10 | Fotonation Cayman Limited | Systems and methods for performing high speed video capture and depth estimation using array cameras |
CN105611123B (en) * | 2015-12-18 | 2017-05-24 | 广东欧珀移动通信有限公司 | imaging method, image sensor, imaging device and electronic device |
CN105578076A (en) * | 2015-12-18 | 2016-05-11 | 广东欧珀移动通信有限公司 | Imaging method, imaging device and electronic device |
CN105592303B (en) * | 2015-12-18 | 2018-09-11 | 广东欧珀移动通信有限公司 | imaging method, imaging device and electronic device |
CN105635702B (en) * | 2015-12-18 | 2017-06-13 | 广东欧珀移动通信有限公司 | Imaging method, imaging device and electronic installation |
CN105578072A (en) * | 2015-12-18 | 2016-05-11 | 广东欧珀移动通信有限公司 | Imaging method, imaging device and electronic device |
CN105578006B (en) * | 2015-12-18 | 2018-02-13 | 广东欧珀移动通信有限公司 | Imaging method, imaging device and electronic installation |
CN105611125B (en) * | 2015-12-18 | 2018-04-10 | 广东欧珀移动通信有限公司 | Imaging method, imaging device and electronic installation |
CN105611257B (en) * | 2015-12-18 | 2018-09-11 | 广东欧珀移动通信有限公司 | Imaging method, imaging sensor, imaging device and electronic device |
CN105578080B (en) * | 2015-12-18 | 2019-02-05 | Oppo广东移动通信有限公司 | Imaging method, imaging device and electronic device |
CN105430361B (en) * | 2015-12-18 | 2018-03-20 | 广东欧珀移动通信有限公司 | Imaging method, imaging sensor, imaging device and electronic installation |
CN105578077A (en) * | 2015-12-18 | 2016-05-11 | 广东欧珀移动通信有限公司 | Imaging method, imaging device and electronic device |
CN105635532B (en) * | 2015-12-18 | 2017-05-24 | 广东欧珀移动通信有限公司 | imaging method, image sensor, imaging device and electronic device |
US10683034B2 (en) * | 2017-06-06 | 2020-06-16 | Ford Global Technologies, Llc | Vehicle remote parking systems and methods |
US10585430B2 (en) | 2017-06-16 | 2020-03-10 | Ford Global Technologies, Llc | Remote park-assist authentication for vehicles |
US10775781B2 (en) | 2017-06-16 | 2020-09-15 | Ford Global Technologies, Llc | Interface verification for vehicle remote park-assist |
US10482618B2 (en) | 2017-08-21 | 2019-11-19 | Fotonation Limited | Systems and methods for hybrid depth regularization |
JP6963295B2 (en) * | 2017-09-01 | 2021-11-05 | 学校法人東京電機大学 | 3D information acquisition device |
US10580304B2 (en) | 2017-10-02 | 2020-03-03 | Ford Global Technologies, Llc | Accelerometer-based external sound monitoring for voice controlled autonomous parking |
US10627811B2 (en) | 2017-11-07 | 2020-04-21 | Ford Global Technologies, Llc | Audio alerts for remote park-assist tethering |
CN107846537B (en) | 2017-11-08 | 2019-11-26 | 维沃移动通信有限公司 | A kind of CCD camera assembly, image acquiring method and mobile terminal |
US10578676B2 (en) | 2017-11-28 | 2020-03-03 | Ford Global Technologies, Llc | Vehicle monitoring of mobile device state-of-charge |
US10583830B2 (en) | 2018-01-02 | 2020-03-10 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10688918B2 (en) | 2018-01-02 | 2020-06-23 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10814864B2 (en) | 2018-01-02 | 2020-10-27 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US11148661B2 (en) | 2018-01-02 | 2021-10-19 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10585431B2 (en) | 2018-01-02 | 2020-03-10 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10737690B2 (en) | 2018-01-02 | 2020-08-11 | Ford Global Technologies, Llc | Mobile device tethering for a remote parking assist system of a vehicle |
US10974717B2 (en) | 2018-01-02 | 2021-04-13 | Ford Global Technologies, I.LC | Mobile device tethering for a remote parking assist system of a vehicle |
US10684773B2 (en) | 2018-01-03 | 2020-06-16 | Ford Global Technologies, Llc | Mobile device interface for trailer backup-assist |
US10747218B2 (en) | 2018-01-12 | 2020-08-18 | Ford Global Technologies, Llc | Mobile device tethering for remote parking assist |
US10917748B2 (en) | 2018-01-25 | 2021-02-09 | Ford Global Technologies, Llc | Mobile device tethering for vehicle systems based on variable time-of-flight and dead reckoning |
US10684627B2 (en) | 2018-02-06 | 2020-06-16 | Ford Global Technologies, Llc | Accelerometer-based external sound monitoring for position aware autonomous parking |
US11188070B2 (en) | 2018-02-19 | 2021-11-30 | Ford Global Technologies, Llc | Mitigating key fob unavailability for remote parking assist systems |
US10507868B2 (en) | 2018-02-22 | 2019-12-17 | Ford Global Technologies, Llc | Tire pressure monitoring for vehicle park-assist |
US10732622B2 (en) | 2018-04-05 | 2020-08-04 | Ford Global Technologies, Llc | Advanced user interaction features for remote park assist |
US10793144B2 (en) | 2018-04-09 | 2020-10-06 | Ford Global Technologies, Llc | Vehicle remote park-assist communication counters |
US10493981B2 (en) | 2018-04-09 | 2019-12-03 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10759417B2 (en) | 2018-04-09 | 2020-09-01 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10683004B2 (en) | 2018-04-09 | 2020-06-16 | Ford Global Technologies, Llc | Input signal management for vehicle park-assist |
US10384605B1 (en) | 2018-09-04 | 2019-08-20 | Ford Global Technologies, Llc | Methods and apparatus to facilitate pedestrian detection during remote-controlled maneuvers |
US10717432B2 (en) | 2018-09-13 | 2020-07-21 | Ford Global Technologies, Llc | Park-assist based on vehicle door open positions |
US10821972B2 (en) | 2018-09-13 | 2020-11-03 | Ford Global Technologies, Llc | Vehicle remote parking assist systems and methods |
US10967851B2 (en) | 2018-09-24 | 2021-04-06 | Ford Global Technologies, Llc | Vehicle system and method for setting variable virtual boundary |
US10529233B1 (en) | 2018-09-24 | 2020-01-07 | Ford Global Technologies Llc | Vehicle and method for detecting a parking space via a drone |
US10908603B2 (en) | 2018-10-08 | 2021-02-02 | Ford Global Technologies, Llc | Methods and apparatus to facilitate remote-controlled maneuvers |
US10628687B1 (en) | 2018-10-12 | 2020-04-21 | Ford Global Technologies, Llc | Parking spot identification for vehicle park-assist |
US11097723B2 (en) | 2018-10-17 | 2021-08-24 | Ford Global Technologies, Llc | User interfaces for vehicle remote park assist |
US11137754B2 (en) | 2018-10-24 | 2021-10-05 | Ford Global Technologies, Llc | Intermittent delay mitigation for remote vehicle operation |
CN109579752B (en) * | 2018-11-20 | 2021-07-23 | 维沃移动通信有限公司 | Measuring method and terminal equipment |
US11789442B2 (en) | 2019-02-07 | 2023-10-17 | Ford Global Technologies, Llc | Anomalous input detection |
US11195344B2 (en) | 2019-03-15 | 2021-12-07 | Ford Global Technologies, Llc | High phone BLE or CPU burden detection and notification |
US11275368B2 (en) | 2019-04-01 | 2022-03-15 | Ford Global Technologies, Llc | Key fobs for vehicle remote park-assist |
US11169517B2 (en) | 2019-04-01 | 2021-11-09 | Ford Global Technologies, Llc | Initiation of vehicle remote park-assist with key fob |
KR102646521B1 (en) | 2019-09-17 | 2024-03-21 | 인트린식 이노베이션 엘엘씨 | Surface modeling system and method using polarization cue |
CN114766003B (en) | 2019-10-07 | 2024-03-26 | 波士顿偏振测定公司 | Systems and methods for enhancing sensor systems and imaging systems with polarization |
CN114787648B (en) | 2019-11-30 | 2023-11-10 | 波士顿偏振测定公司 | Systems and methods for transparent object segmentation using polarization cues |
US11470287B2 (en) | 2019-12-05 | 2022-10-11 | Samsung Electronics Co., Ltd. | Color imaging apparatus using monochrome sensors for mobile devices |
KR20210081767A (en) * | 2019-12-24 | 2021-07-02 | 삼성전자주식회사 | Imaging device and image sensing method |
CN115552486A (en) | 2020-01-29 | 2022-12-30 | 因思创新有限责任公司 | System and method for characterizing an object pose detection and measurement system |
WO2021154459A1 (en) | 2020-01-30 | 2021-08-05 | Boston Polarimetrics, Inc. | Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images |
JP2023539884A (en) * | 2020-08-31 | 2023-09-20 | トライアイ リミテッド | Shortwave infrared focal plane array, and methods for its use and manufacturing |
US11290658B1 (en) | 2021-04-15 | 2022-03-29 | Boston Polarimetrics, Inc. | Systems and methods for camera exposure control |
US11689813B2 (en) | 2021-07-01 | 2023-06-27 | Intrinsic Innovation Llc | Systems and methods for high dynamic range imaging using crossed polarizers |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5570221A (en) * | 1993-12-28 | 1996-10-29 | Nec Corporation | Light amplification device |
US5729011A (en) * | 1995-02-24 | 1998-03-17 | Olympus Optical Co., Ltd. | Spectroscopic apparatus and spectroscopic image recording apparatus |
US20020090650A1 (en) * | 2000-04-06 | 2002-07-11 | Quantum Dot Corporation | Two-dimensional spectral imaging system |
US6639664B2 (en) * | 1998-05-13 | 2003-10-28 | Storz Endoskop Gmbh | Endoscope for inspection of an observation cavity |
US20050057491A1 (en) * | 2003-08-28 | 2005-03-17 | Eastman Kodak Company | Private display system |
US6937279B1 (en) * | 1998-12-22 | 2005-08-30 | Hynix Semiconductor Inc. | Apparatus for converting analog image data into digital image data in CMOS image sensor |
US6961080B2 (en) * | 1997-04-09 | 2005-11-01 | Richardson Technologies Inc. | Color translating UV microscope |
US20070034777A1 (en) * | 2005-08-12 | 2007-02-15 | Tessera, Inc. | Image sensor employing a plurality of photodetector arrays and/or rear-illuminated architecture |
US7242478B1 (en) * | 2003-12-05 | 2007-07-10 | Surface Optics Corporation | Spatially corrected full-cubed hyperspectral imager |
US20070260413A1 (en) * | 2004-10-28 | 2007-11-08 | Peter Ehbets | Method for Correcting Measured Image Values |
US20080204744A1 (en) * | 2005-07-11 | 2008-08-28 | Jose Mir | High Speed, Optically-Multiplexed, Hyperspectral Imagers and Methods Thereof |
US7605958B2 (en) * | 2004-06-02 | 2009-10-20 | Xerox Corporation | Design parameters for a multi-row linear photosensor array |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6650361B1 (en) * | 1997-12-17 | 2003-11-18 | Canon Kabushiki Kaisha | Imaging apparatus control method, and a computer program product having computer program code therefor |
JP3451195B2 (en) * | 1998-05-28 | 2003-09-29 | 富士写真フイルム株式会社 | Image reading device |
US6069973A (en) * | 1998-06-30 | 2000-05-30 | Xerox Corporation | Method and apparatus for color correction in a multi-chip imaging array |
US6667818B1 (en) * | 2000-10-23 | 2003-12-23 | Umax Data Systems, Inc. | Multiple-field sensor for scanning system |
JP3703424B2 (en) * | 2000-10-25 | 2005-10-05 | キヤノン株式会社 | IMAGING DEVICE, ITS CONTROL METHOD, CONTROL PROGRAM, AND STORAGE MEDIUM |
JP2002135796A (en) * | 2000-10-25 | 2002-05-10 | Canon Inc | Imaging apparatus |
JP2003143459A (en) * | 2001-11-02 | 2003-05-16 | Canon Inc | Compound-eye image pickup system and device provided therewith |
JP2003283907A (en) * | 2002-03-20 | 2003-10-03 | Japan Science & Technology Corp | Imaging device |
EP2466871A3 (en) | 2003-10-22 | 2017-05-03 | Panasonic Intellectual Property Management Co., Ltd. | Imaging apparatus and method for producing the same, portable equipment, and imaging sensor and method for producing the same. |
JP2005176040A (en) * | 2003-12-12 | 2005-06-30 | Canon Inc | Imaging device |
US7511749B2 (en) | 2003-12-18 | 2009-03-31 | Aptina Imaging Corporation | Color image sensor having imaging element array forming images on respective regions of sensor elements |
US7773143B2 (en) | 2004-04-08 | 2010-08-10 | Tessera North America, Inc. | Thin color camera having sub-pixel resolution |
US7830426B2 (en) * | 2005-12-29 | 2010-11-09 | Micron Technology, Inc. | Method and apparatus providing color interpolation in color filter arrays using edge detection and correction terms |
-
2006
- 2006-03-06 US US11/367,580 patent/US7924483B2/en active Active
- 2006-12-21 US US11/642,867 patent/US20070206242A1/en not_active Abandoned
-
2007
- 2007-03-02 WO PCT/US2007/005439 patent/WO2007103212A1/en active Application Filing
- 2007-03-02 CN CNA2007800078540A patent/CN101395926A/en active Pending
- 2007-03-02 KR KR1020087023960A patent/KR20080109794A/en not_active Application Discontinuation
- 2007-03-02 JP JP2008558320A patent/JP2009529291A/en active Pending
- 2007-03-02 EP EP07752157A patent/EP2008470A1/en not_active Withdrawn
- 2007-03-06 TW TW096107734A patent/TWI398948B/en not_active IP Right Cessation
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5570221A (en) * | 1993-12-28 | 1996-10-29 | Nec Corporation | Light amplification device |
US5729011A (en) * | 1995-02-24 | 1998-03-17 | Olympus Optical Co., Ltd. | Spectroscopic apparatus and spectroscopic image recording apparatus |
US6961080B2 (en) * | 1997-04-09 | 2005-11-01 | Richardson Technologies Inc. | Color translating UV microscope |
US6639664B2 (en) * | 1998-05-13 | 2003-10-28 | Storz Endoskop Gmbh | Endoscope for inspection of an observation cavity |
US6937279B1 (en) * | 1998-12-22 | 2005-08-30 | Hynix Semiconductor Inc. | Apparatus for converting analog image data into digital image data in CMOS image sensor |
US20020090650A1 (en) * | 2000-04-06 | 2002-07-11 | Quantum Dot Corporation | Two-dimensional spectral imaging system |
US20050057491A1 (en) * | 2003-08-28 | 2005-03-17 | Eastman Kodak Company | Private display system |
US7242478B1 (en) * | 2003-12-05 | 2007-07-10 | Surface Optics Corporation | Spatially corrected full-cubed hyperspectral imager |
US7605958B2 (en) * | 2004-06-02 | 2009-10-20 | Xerox Corporation | Design parameters for a multi-row linear photosensor array |
US20070260413A1 (en) * | 2004-10-28 | 2007-11-08 | Peter Ehbets | Method for Correcting Measured Image Values |
US20080204744A1 (en) * | 2005-07-11 | 2008-08-28 | Jose Mir | High Speed, Optically-Multiplexed, Hyperspectral Imagers and Methods Thereof |
US20070034777A1 (en) * | 2005-08-12 | 2007-02-15 | Tessera, Inc. | Image sensor employing a plurality of photodetector arrays and/or rear-illuminated architecture |
Cited By (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070201114A1 (en) * | 2006-02-28 | 2007-08-30 | Yoshitaka Egawa | Solid-state image sensing device having photoelectric conversion cells each configured by n pixels in vertical direction |
US8854420B2 (en) | 2007-12-27 | 2014-10-07 | Google Inc. | High-resolution, variable depth of field image device |
WO2009085305A1 (en) | 2007-12-27 | 2009-07-09 | Google Inc. | High-resolution, variable depth of field image device |
EP2238745A1 (en) * | 2007-12-27 | 2010-10-13 | Google, Inc. | High-resolution, variable depth of field image device |
US20110134224A1 (en) * | 2007-12-27 | 2011-06-09 | Google Inc. | High-Resolution, Variable Depth of Field Image Device |
EP2238745A4 (en) * | 2007-12-27 | 2012-02-22 | Google Inc | High-resolution, variable depth of field image device |
US8319822B2 (en) | 2007-12-27 | 2012-11-27 | Google Inc. | High-resolution, variable depth of field image device |
US9217671B2 (en) | 2009-09-01 | 2015-12-22 | Koninklijke Philips N.V. | High spectral resolution color sensor using non-dispersive elements |
US8558933B2 (en) * | 2010-12-09 | 2013-10-15 | Sony Corporation | Imaging device and imaging apparatus |
US20120147233A1 (en) * | 2010-12-09 | 2012-06-14 | Sony Corporation | Imaging device and imaging apparatus |
US20140267878A1 (en) * | 2011-11-04 | 2014-09-18 | Imec | Spectral camera with overlapping segments of image copies interleaved onto sensor array |
US9772229B2 (en) | 2011-11-04 | 2017-09-26 | Imec | Spectral camera with integrated filters and multiple adjacent image copies projected onto sensor array |
US9366573B2 (en) * | 2011-11-04 | 2016-06-14 | Imec Leuven | Spectral camera with overlapping segments of image copies interleaved onto sensor array |
US20150077524A1 (en) * | 2012-03-16 | 2015-03-19 | Nikon Corporation | Image sensor and imaging device |
US10018758B2 (en) * | 2012-06-05 | 2018-07-10 | Hypermed Imaging, Inc. | Single-sensor hyperspectral imaging device |
US11092725B2 (en) | 2012-06-05 | 2021-08-17 | Samsung Electronics Co., Ltd. | Single-sensor hyperspectral imaging device |
US9766382B2 (en) * | 2012-06-05 | 2017-09-19 | Hypermed Imaging, Inc. | Single-sensor hyperspectral imaging device |
US20150177429A1 (en) * | 2012-06-05 | 2015-06-25 | Hypermed Imaging, Inc. | Single-sensor hyperspectral imaging device |
US20180275325A1 (en) * | 2012-06-05 | 2018-09-27 | Hypermed Imaging, Inc. | Single-sensor hyperspectral imaging device |
US10534116B2 (en) * | 2012-06-05 | 2020-01-14 | Hypermed Imaging, Inc. | Single-sensor hyperspectral imaging device |
US11493675B2 (en) | 2012-06-05 | 2022-11-08 | Samsung Electronics Co., Ltd. | Single-sensor hyperspectral imaging device |
US9392237B2 (en) * | 2013-05-21 | 2016-07-12 | Olympus Corporation | Image processing device and image processing method |
US20140347533A1 (en) * | 2013-05-21 | 2014-11-27 | Olympus Corporation | Image processing device and image processing method |
US10186540B2 (en) * | 2013-07-30 | 2019-01-22 | Heptagon Micro Optics Pte. Ltd. | Optoelectronic modules that have shielding to reduce light leakage or stray light |
US20150325613A1 (en) * | 2013-07-30 | 2015-11-12 | Heptagon Micro Optics Pte. Ltd. | Optoelectronic modules that have shielding to reduce light leakage or stray light, and fabrication methods for such modules |
US20190190237A1 (en) * | 2016-09-29 | 2019-06-20 | Intel Corporation | Optical free air bus interconnect |
CN110235252A (en) * | 2016-12-07 | 2019-09-13 | 兴业法兰西红外探测器-索弗拉迪尔公司 | For improving the device of multispectral detection |
US20200329205A1 (en) * | 2020-04-02 | 2020-10-15 | Sichuan Dualix Spectral Lmaging Technology Co., Ltd | Hyperspectral camera based on continuously variable film filter and coating method thereof |
US11622084B2 (en) * | 2020-04-02 | 2023-04-04 | Jiangsu Dualix Spectral Imaging Technology Co., Ltd | Hyperspectral camera based on continuously variable film filter and coating method thereof |
US20220214217A1 (en) * | 2021-01-04 | 2022-07-07 | Argo AI, LLC | Systems and methods for characterizing spectral reflectance of real world objects |
US11536608B2 (en) * | 2021-01-04 | 2022-12-27 | Argo AI, LLC | Systems and methods for characterizing spectral reflectance of real world objects |
Also Published As
Publication number | Publication date |
---|---|
KR20080109794A (en) | 2008-12-17 |
TW200742054A (en) | 2007-11-01 |
US20070206241A1 (en) | 2007-09-06 |
WO2007103212A1 (en) | 2007-09-13 |
JP2009529291A (en) | 2009-08-13 |
TWI398948B (en) | 2013-06-11 |
US7924483B2 (en) | 2011-04-12 |
EP2008470A1 (en) | 2008-12-31 |
CN101395926A (en) | 2009-03-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20070206242A1 (en) | Method, apparatus and system providing an integrated hyperspectral imager | |
US11916093B2 (en) | Solid-state imaging device, driving method therefor, and electronic apparatus | |
US10297629B2 (en) | Image sensors with in-pixel lens arrays | |
US7242478B1 (en) | Spatially corrected full-cubed hyperspectral imager | |
US7433042B1 (en) | Spatially corrected full-cubed hyperspectral imager | |
KR101890940B1 (en) | Imaging device and imaging apparatus | |
TWI395324B (en) | Method, apparatus and system providing holographic layer as micro-lens and color filter array in an imager | |
US20080080028A1 (en) | Imaging method, apparatus and system having extended depth of field | |
US9666631B2 (en) | Photodiode and filter configuration for high dynamic range image sensor | |
US20080170228A1 (en) | Method and apparatus for wafer level calibration of imaging sensors | |
US6661457B1 (en) | Pixel read-out architecture | |
US20120062774A1 (en) | Adaptive solid state image sensor | |
JP6348271B2 (en) | Mixed material multispectral Stirling array sensor | |
KR20080091023A (en) | Solid-state imaging device and imaging apparatus | |
US10165211B1 (en) | Image sensors with optically black pixels | |
US10942304B2 (en) | Solid-state imaging element, manufacturing method of the same, and electronic device | |
US9921106B1 (en) | Integrated imaging spectrometer for hyperspectral imaging systems | |
US8416327B2 (en) | Solid-state image pickup apparatus | |
US8077230B2 (en) | Methods and apparatus for reducing color material related defects in imagers | |
US20050109917A1 (en) | Multi-spectral imaging with almost-full fill-factor using 3D pixels | |
CN113824906A (en) | Image sensing device | |
KR20220041351A (en) | Image sensing device | |
US20210266431A1 (en) | Imaging sensor pixels having built-in grating | |
US11817468B2 (en) | Image sensing device | |
US11917272B2 (en) | Imaging systems for multi-spectral imaging |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICRON TECHNOLOGY, INC., IDAHO Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SMITH, SCOTT;REEL/FRAME:018717/0427 Effective date: 20061218 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |