US20070206242A1 - Method, apparatus and system providing an integrated hyperspectral imager - Google Patents

Method, apparatus and system providing an integrated hyperspectral imager Download PDF

Info

Publication number
US20070206242A1
US20070206242A1 US11/642,867 US64286706A US2007206242A1 US 20070206242 A1 US20070206242 A1 US 20070206242A1 US 64286706 A US64286706 A US 64286706A US 2007206242 A1 US2007206242 A1 US 2007206242A1
Authority
US
United States
Prior art keywords
hyperspectral
imager apparatus
bandpass
pixel arrays
imager
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/642,867
Inventor
Scott Smith
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Micron Technology Inc
Original Assignee
Micron Technology Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Micron Technology Inc filed Critical Micron Technology Inc
Priority to US11/642,867 priority Critical patent/US20070206242A1/en
Assigned to MICRON TECHNOLOGY, INC. reassignment MICRON TECHNOLOGY, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SMITH, SCOTT
Publication of US20070206242A1 publication Critical patent/US20070206242A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/1462Coatings
    • H01L27/14621Colour filter arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/10Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
    • H04N23/13Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths with multiple sensors
    • H04N23/15Image signal generation with circuitry for avoiding or correcting image misregistration
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/1462Coatings
    • H01L27/14623Optical shielding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/40Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
    • H04N25/41Extracting pixel data from a plurality of image sensors simultaneously picking up an image, e.g. for increasing the field of view by combining the outputs of a plurality of sensors

Definitions

  • Embodiments of the invention relate generally to the field of semiconductor devices and more particularly to multi-array image sensors.
  • Image sensors such as multispectral image sensors, generally produce images with a few relatively broad wavelength bands from a wavelength of about 400 nm to about 700 nm. These bands typically correlate to the red, green and blue color filters (RGB) of a Bayer patterned color filter array (described below) used in the image sensor.
  • RGB red, green and blue color filters
  • Hyperspectral image sensors simultaneously collect image data in dozens or hundreds of narrow, adjacent hyperspectral bands. Hyperspectral sensors create a larger number of images from contiguous, rather than disjoint, regions of the spectrum, typically, with much finer resolution than can be obtained with a multispectral image sensor. Hyperspectral imaging involves acquisition of image data in many contiguous narrow hyperspectral bands, the goal being to produce laboratory quality reflectance spectra for each pixel in an image.
  • CCDs charge coupled devices
  • APS CMOS active pixel sensors
  • photodiode arrays charge injection devices
  • hybrid focal plane arrays among others.
  • CCDs charge coupled devices
  • photodiode arrays charge injection devices
  • hybrid focal plane arrays among others.
  • image sensors use imaging lenses to focus electromagnetic radiation onto photo-conversion devices, e.g., photodiodes.
  • these image sensors can use color filters to pass particular wavelengths of electromagnetic radiation to the photo-conversion devices, such that the photo-conversion devices typically are associated with a particular color.
  • FIGS. 1A and 1B respectively show a top view and a simplified cross sectional view of a portion of a conventional color image sensor for the visual light spectrum using a Bayer patterned color filter pixel cell array 100 .
  • the array 100 includes pixel cells 10 , each being formed on a substrate 1 .
  • the pixel array 100 is covered by a color filter array 30 .
  • the color filter array 30 includes color filters 31 r, 31 g, 31 b, each disposed over a respective pixel cell 10 .
  • Each of the filters 31 r, 31 g, 31 b allows only particular wavelengths of light to pass through to a respective photo-conversion device.
  • each pixel cell 10 includes a photo-conversion device 12 r, 12 g, for example, a photodiode, having an associated charge collecting well 13 r, 13 g.
  • the elements 12 r, 12 g, 13 r, 13 g are associated with red and green colors based on the color being passed to the pixels of one row of the color filter array 30 , however, it should be appreciated that there may be a photo-conversion device 12 and a charge collecting well 13 that is associated with the color blue that is not shown in the cross sectional view of FIG. 1B .
  • the illustrated array 100 has imaging lenses 20 that collect and focus light onto respective photo-conversion devices 12 r, 12 g, which in turn convert the focused light into electrons that are stored in the respective charge collecting wells 13 r, 13 g.
  • a passivation layer 6 which typically covers the gate structure of transistors of the pixels and an overlying interlayer dielectric (ILD) region 3 .
  • the ILD region 3 typically includes multiple layers of interlayer dielectrics and conductors that form connections between devices of the pixel cells 10 and from the pixel cells 10 to circuitry 150 peripheral to the array 100 .
  • a dielectric layer 5 may also be provided between the color filter array 30 and imaging lenses 20 .
  • a hyperspectral image sensor or camera system relies on many narrow hyperspectral bandpass filters to capture the hyperspectral image content of a scene.
  • the hyperspectral bandpass filters may be applied to an imaging system of the type illustrated in FIG. 1B .
  • One major disadvantage of this is that because of the large number of narrow band filters required, the hyperspectral image sensor becomes very large and expensive. Further, because the image sensor relies on applying separate hyperspectral bandpass filters to a lens system over one or more pixels in each pixel array, the filters do not capture the full image content and may produce poor quality images.
  • FIG. 1A is a top plan view of a portion of a conventional Bayer pattern color image sensor.
  • FIG. 1B is a cross sectional view of a row portion of a conventional Bayer pattern color image sensor.
  • FIG. 2A is a plan view of a multi-array image sensor in accordance with an embodiment described herein.
  • FIG. 2B is a cross sectional view along a line A-A of the FIG. 2A image sensor in accordance with an embodiment described herein.
  • FIG. 3 is a graph showing the measurements of hyperspectral imaging in accordance with an embodiment described herein.
  • FIG. 4 is a graph showing the measurements of a continuous spectrum for each hyperspectral image from each pixel in accordance with an embodiment described herein.
  • FIG. 5 is a cross sectional view of a portion of an image sensor in accordance to an embodiment described herein.
  • FIG. 6 illustrates a block diagram of a CMOS image sensor constructed in accordance with an embodiment described herein.
  • FIG. 7 depicts a processor system constructed in accordance with an embodiment described herein.
  • the embodiments described herein relate to methods, apparatuses, and systems for integrating a plurality of pixel arrays onto a single substrate, each having an associated hyperspectral bandpass filter and imaging lens.
  • Each imaging array and its associated lens and bandpass filter form an image of objects in a scene. If the image sensor arrays and lenses are in close proximity, on the order of less than 2 mm, e.g., about 0.5 mm to about 1 mm, then object parallax will be quite small and each array shared will image the same scene for objects that are all farther than 1 m from the arrays.
  • a hyperspectral image may be constructed from individual images respectively acquired by the pixel arrays such that each pixel in the image contains a hyperspectral representation of reflectance of that point in the scene.
  • FIG. 2A is a plan view of an embodiment of a hyperspectral image sensor 200 on a semiconductor, e.g., silicon, substrate 201 of a single die.
  • image sensor 200 is formed as a 4 ⁇ 4 array configuration of pixel arrays 203 for a total of 16 pixel arrays 203 , each having an associated lens element 204 .
  • the imaging lenses 204 are formed by molding or by etching a lens material layer.
  • a hyperspectral bandpass filter 205 ( FIG. 2B ) may be formed over each optical imaging lens 204 .
  • the hyperspectral bandpass filters 205 may be formed separate from lenses 204 , as illustrated in FIG.
  • Each hyperspectral bandpass filter 205 associated with an imaging lens 204 has a different narrow hyperspectral bandpass characteristic.
  • there are sixteen different narrow hyperspectral bandpass filters 205 each passing a narrow wavelength range of light to a respective pixel array 203 below.
  • Hyperspectral bandpass filters 205 may pass narrow bandwidths that are narrower than the bandwidths used to simply separate red, green and blue wavelengths and may have pass bands of and provide a spectral resolution of less than 100 nm per band, e.g., 50-100 nm per band, and could include infrared bands from about 700 nm to about 1000 nm and/or UV bands from about 200 nm to about 400 nm.
  • the filters 205 may also have a bandwidth resolution that is as narrow as less than 10 nm per band, for example, 5-10 nm per band.
  • image sensor 200 includes a substrate 201 on which a plurality of pixel arrays 203 and associated support circuitry (not shown) for each array 203 are fabricated.
  • the illustrated portion shows one pixel array 203 for one color that has a corresponding color filter that only allows a limited narrow spectrum of wavelengths to pass through.
  • Each imaging lens 204 or pixel array 203 then captures different narrow band hyperspectral information.
  • the embodiment shown in FIG. 2B shows pixel arrays 203 that are four pixels wide, but it should be appreciated that each of the pixel arrays 203 may contain hundreds, thousands or millions of pixels, as desired.
  • Opaque walls 260 separate the individual arrays 203 and opaque walls 270 separate the arrays of imaging lenses 204 that are arranged above each respective pixel array 203 . Opaque walls may also optionally separate the narrow hyperspectral filters 205 .
  • the hyperspectral bandpass filters 205 are unique for each imaging lens 204 , meaning that each imaging lens 204 or pixel array 203 under each lens 204 has a different narrow band color filter that only allows a limited spectrum of wavelengths to pass through. Thus, each pixel array 203 captures different narrow band hyperspectral information.
  • the pixel arrays 203 preferably are integrated on a single silicon die substrate with common circuitry. Including a multi-array color image sensor on a single die provides for a reduction of color crosstalk artifacts, especially for compact camera modules with pixel sizes less than 6 microns by 6 microns. Moreover, an imaging lens with a short focal length can minimize parallax effects and allow a camera module employing image sensor 200 to be more compact.
  • FIG. 2B illustrates a single lens 204 over each array 203
  • a modified embodiment provides an imaging lens 204 over one or more pixels in each array 203 .
  • each individual lens 204 may cover and focus light on a sixteen pixel section (in a 4 ⁇ 4 pattern) of the pixels in each array 203 .
  • the image sensor 200 employs hyperspectral imaging, the image data is simultaneously collected in dozens or hundreds of narrow, adjacent hyperspectral bands, as illustrated in FIG. 3 .
  • a 4 ⁇ 4 arrangement of pixel arrays (as illustrated in FIGS. 2A-2B ) has at most 16 hyperspectral bands.
  • the image sensor must have a hundred or more individual pixel arrays, e.g., a 10 ⁇ 10 arrangement of pixel arrays.
  • Each image I 1 . . . I n from a respective array represents a hyperspectral image from a respective pixel array 203 ( FIG. 2A ).
  • the totality of images I 1 . . . I n represent an image with all hyperspectral bands, as illustrated in FIG. 4 .
  • the individual images can then be combined and thus, each pixel in the combined image would contain full hyperspectral information.
  • ⁇ 1 . . . ⁇ n represents the bandpass optical filter for each image I 1 . . . I n of FIG. 4 .
  • Each imaging lens 204 with its unique hyperspectral bandpass filter will image the same scene if the image sensor arrays 203 are in close proximity to one another, e.g., on the order of less than 2 mm and with a preferred spacing in the range of about 0.5 mm to about 1 mm.
  • the lenses are similarly spaced.
  • a hyperspectral or full spectral image may be constructed such that each pixel in the produced image contains a full spectral representation of reflectance of that point in the scene.
  • object parallax is quite small.
  • FIG. 5 is a simplified cross sectional view of a portion of an image sensor 200 according to any embodiment of the invention.
  • Sensor 200 includes a substrate 201 on which a plurality of pixel arrays 203 and associated support circuitry (not shown) for each array are fabricated.
  • Each color pixel array 203 is associated with a different color on a wavelength spectrum.
  • each imaging lens 204 has a unique hyperspectral bandpass filter 205 that is associated with each pixel array 203 which is designed to pass a limited spectrum of wavelengths.
  • Each pixel array 203 then captures different hyperspectral information.
  • N is the pixel array dimension (in pixels)
  • d x is the distance between the outer focal points of lenses 204 and where x is approximated by the following formula:
  • O is the object distance from the camera (or distance from the object to be imaged and the imaging lenses 204 ) and ⁇ is the field of view (FOV) angle.
  • FOV field of view
  • the P x or parallax distance at the object plane between adjacent pixel arrays 203 would equal 0.78 pixels by using formulas (1) and (2) above.
  • the maximum parallax (at 60° FOV) between images for objects with a distance of greater than or equal to 2 m from the camera would equal 2.35 pixels.
  • a 4 ⁇ 4 image sensor would have 16 hyperspectral bands, a pixel element size of 2.2 ⁇ m ⁇ 2.2 ⁇ m and a target hyperspectral image size of 800 ⁇ 600.
  • Various image sensor configurations allow tradeoffs in performance. For example, a 3 ⁇ 3 image sensor will have less maximum parallax than a 4 ⁇ 4 image sensor, but will have only 9 hyperspectral bands. A larger FOV will have less parallax but the user must get closer to objects they want to isolate in a scene. Further, smaller pixels may be used to reduce maximum parallax with a tradeoff of a reduction in camera sensitivity.
  • FIG. 6 illustrates a block diagram of a CMOS imaging device 500 , which employs a multi-array sensor 200 having a plurality of pixel arrays 203 constructed according to any embodiment described above. Multiple arrays 203 are arranged to form one large pixel array 505 . The pixels of each row in array 505 are all turned on at the same time by a row select line, and the pixels of each column are selectively output onto the output lines by respective column select lines. A plurality of row and column select lines are provided for the entire array 505 . The row lines are selectively activated in sequence by a row driver 510 in response to row address decoder 515 . The column select lines are selectively activated in sequence for each row activation by a column driver 520 in response to column address decoder 525 . Thus, a row and column address is provided for each pixel.
  • the CMOS image sensor 500 is operated by a control circuit 530 , which controls address decoders 515 , 525 for selecting the appropriate row and column select lines for pixel readout.
  • Control circuit 530 also controls the row and column driver circuitry 510 , 520 so that they apply driving voltages to the drive transistors of the selected row and column select lines.
  • the pixel output signals typically include a pixel reset signal V rst read out of the storage region after it is reset by the reset transistor and a pixel image signal V sig , which is read out of the storage region after photo-generated charges are transferred to the region.
  • V rst and V sig signals are sampled by a sample and hold circuit 535 and are subtracted by a differential amplifier 540 , to produce a differential signal V rst -V sig for each readout pixel.
  • V rst -V sig represents the amount of light impinging on the pixels.
  • This difference signal is digitized by an analog-to-digital converter 545 .
  • the digitized pixel signals are fed to an image processor 550 to form a digital image output.
  • the digitizing and image processing can be located on or off the imaging device chip.
  • the differential signal V rst -V sig can be amplified as a differential signal and directly digitized by a differential analog to digital converter.
  • the analog-to digital converter 545 supplies the digitized pixel signals to an image processor 550 , which performs appropriate image processing, which can include combining the outputs of multiple arrays and performing a parallax adjustment calculation if needed or desired, before providing digital signals defining an image output.
  • FIG. 6 represents one example of a readout circuit for multi-array 505 .
  • Another embodiment could employ the readout circuit in FIG. 6 for each individual array 203 with an image processor 550 combining the outputs of the individual arrays 203 .
  • FIG. 6 illustrates a readout circuit suitable for CMOS arrays 203 , but the invention may be used with other solid state imaging arrays, for example, CCD arrays in which case a readout circuit suitable for reading out CCD arrays could be employed.
  • FIG. 7 illustrates a processor system 900 including the image sensor 500 of FIG. 6 .
  • the processor system 900 is one example of a system having digital circuits that could include image sensor devices. Without being limiting, such a processor system could also include a computer system, camera system, scanner, machine vision, vehicle navigation, video phone, surveillance system, auto focus system, star tracker system, motion detection system, image stabilization system, and other processor system.
  • the processor system 900 for example a digital camera system, generally comprises a central processing unit (CPU) 995 , such as a microprocessor for common operational control, that communicates with an input/output (I/O) device 991 over a bus 993 .
  • Image sensor 500 also communicates with the CPU 995 over the bus 993 .
  • the processor-based system 900 also includes random access memory (RAM) 992 , and can include removable memory 994 , such as flash memory, which also communicate with CPU 995 over the bus 993 .
  • Image sensor 500 may be combined with a processor, such as a CPU, digital signal processor, or microprocessor, with or without memory storage on a single integrated circuit or on a different chip than the processor.
  • a parallax adjustment calculation may be performed by the image processor 550 in image sensor 500 , or by the CPU 995 .

Abstract

Methods, apparatuses, and systems are disclosed which provide a plurality of pixel arrays on a common substrate each associated with a hyperspectral filter. Images from each of the arrays may be separately analyzed or combined into a composite image.

Description

    CROSS REFERENCE TO RELATED APPLICATION
  • This application is a continuation-in-part of U.S. application Ser. No. 11/367,580 filed Mar. 6, 2006, the disclosure of which is incorporated by reference herein.
  • FIELD OF THE INVENTION
  • Embodiments of the invention relate generally to the field of semiconductor devices and more particularly to multi-array image sensors.
  • BACKGROUND OF THE INVENTION
  • Image sensors, such as multispectral image sensors, generally produce images with a few relatively broad wavelength bands from a wavelength of about 400 nm to about 700 nm. These bands typically correlate to the red, green and blue color filters (RGB) of a Bayer patterned color filter array (described below) used in the image sensor. Hyperspectral image sensors, on the other hand, simultaneously collect image data in dozens or hundreds of narrow, adjacent hyperspectral bands. Hyperspectral sensors create a larger number of images from contiguous, rather than disjoint, regions of the spectrum, typically, with much finer resolution than can be obtained with a multispectral image sensor. Hyperspectral imaging involves acquisition of image data in many contiguous narrow hyperspectral bands, the goal being to produce laboratory quality reflectance spectra for each pixel in an image.
  • The semiconductor industry currently produces different types of semiconductor-based image sensors, such as charge coupled devices (CCDs), CMOS active pixel sensors (APS), photodiode arrays, charge injection devices and hybrid focal plane arrays, among others. These image sensors use imaging lenses to focus electromagnetic radiation onto photo-conversion devices, e.g., photodiodes. Also, these image sensors can use color filters to pass particular wavelengths of electromagnetic radiation to the photo-conversion devices, such that the photo-conversion devices typically are associated with a particular color.
  • FIGS. 1A and 1B respectively show a top view and a simplified cross sectional view of a portion of a conventional color image sensor for the visual light spectrum using a Bayer patterned color filter pixel cell array 100. The array 100 includes pixel cells 10, each being formed on a substrate 1. The pixel array 100 is covered by a color filter array 30. The color filter array 30 includes color filters 31 r, 31 g, 31 b, each disposed over a respective pixel cell 10. Each of the filters 31 r, 31 g, 31 b allows only particular wavelengths of light to pass through to a respective photo-conversion device. Typically, the color filter array 30 is arranged in a repeating Bayer pattern that includes two green color filters 31 g for every red color filter 31 r and blue color filter 31 b, arranged as shown in FIG. 1A. As shown in FIG. 1B, each pixel cell 10 includes a photo- conversion device 12 r, 12 g, for example, a photodiode, having an associated charge collecting well 13 r, 13 g. The elements 12 r, 12 g, 13 r, 13 g are associated with red and green colors based on the color being passed to the pixels of one row of the color filter array 30, however, it should be appreciated that there may be a photo-conversion device 12 and a charge collecting well 13 that is associated with the color blue that is not shown in the cross sectional view of FIG. 1B. The illustrated array 100 has imaging lenses 20 that collect and focus light onto respective photo- conversion devices 12 r, 12 g, which in turn convert the focused light into electrons that are stored in the respective charge collecting wells 13 r, 13 g.
  • Between the color filter array 30 and the pixel cells 10 is a passivation layer 6 which typically covers the gate structure of transistors of the pixels and an overlying interlayer dielectric (ILD) region 3. The ILD region 3 typically includes multiple layers of interlayer dielectrics and conductors that form connections between devices of the pixel cells 10 and from the pixel cells 10 to circuitry 150 peripheral to the array 100. A dielectric layer 5 may also be provided between the color filter array 30 and imaging lenses 20.
  • As discussed above, a hyperspectral image sensor or camera system relies on many narrow hyperspectral bandpass filters to capture the hyperspectral image content of a scene. The hyperspectral bandpass filters may be applied to an imaging system of the type illustrated in FIG. 1B. One major disadvantage of this is that because of the large number of narrow band filters required, the hyperspectral image sensor becomes very large and expensive. Further, because the image sensor relies on applying separate hyperspectral bandpass filters to a lens system over one or more pixels in each pixel array, the filters do not capture the full image content and may produce poor quality images.
  • Therefore, it would be advantageous to have an integrated hyperspectral image sensor which better captures the hyperspectral image content of a scene, which is also compact.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1A is a top plan view of a portion of a conventional Bayer pattern color image sensor.
  • FIG. 1B is a cross sectional view of a row portion of a conventional Bayer pattern color image sensor.
  • FIG. 2A is a plan view of a multi-array image sensor in accordance with an embodiment described herein.
  • FIG. 2B is a cross sectional view along a line A-A of the FIG. 2A image sensor in accordance with an embodiment described herein.
  • FIG. 3 is a graph showing the measurements of hyperspectral imaging in accordance with an embodiment described herein.
  • FIG. 4 is a graph showing the measurements of a continuous spectrum for each hyperspectral image from each pixel in accordance with an embodiment described herein.
  • FIG. 5 is a cross sectional view of a portion of an image sensor in accordance to an embodiment described herein.
  • FIG. 6 illustrates a block diagram of a CMOS image sensor constructed in accordance with an embodiment described herein.
  • FIG. 7 depicts a processor system constructed in accordance with an embodiment described herein.
  • DETAILED DESCRIPTION OF THE INVENTION
  • In the following detailed description, reference is made to the accompanying drawings which form a part hereof, and in which are shown by way of illustration of specific embodiments in which the invention may be practiced. These embodiments are described in sufficient detail to enable those of ordinary skill in the art to make and use them, and it is to be understood that structural, logical or procedural changes may be made to the specific embodiments disclosed herein.
  • The embodiments described herein relate to methods, apparatuses, and systems for integrating a plurality of pixel arrays onto a single substrate, each having an associated hyperspectral bandpass filter and imaging lens. Each imaging array and its associated lens and bandpass filter form an image of objects in a scene. If the image sensor arrays and lenses are in close proximity, on the order of less than 2 mm, e.g., about 0.5 mm to about 1 mm, then object parallax will be quite small and each array shared will image the same scene for objects that are all farther than 1 m from the arrays. A hyperspectral image may be constructed from individual images respectively acquired by the pixel arrays such that each pixel in the image contains a hyperspectral representation of reflectance of that point in the scene.
  • FIG. 2A is a plan view of an embodiment of a hyperspectral image sensor 200 on a semiconductor, e.g., silicon, substrate 201 of a single die. In the illustrated embodiment, image sensor 200 is formed as a 4×4 array configuration of pixel arrays 203 for a total of 16 pixel arrays 203, each having an associated lens element 204. The imaging lenses 204 are formed by molding or by etching a lens material layer. A hyperspectral bandpass filter 205 (FIG. 2B) may be formed over each optical imaging lens 204. The hyperspectral bandpass filters 205 may be formed separate from lenses 204, as illustrated in FIG. 2B, and may be formed above or below the lenses 204, or may be formed as a coating on the lenses 204, or may be formed as part of the lens 204 by using a colored material for lens 204. Each hyperspectral bandpass filter 205 associated with an imaging lens 204 has a different narrow hyperspectral bandpass characteristic. Thus, in the embodiment illustrated in FIGS. 2A-2B, there are sixteen different narrow hyperspectral bandpass filters 205, each passing a narrow wavelength range of light to a respective pixel array 203 below. Hyperspectral bandpass filters 205 may pass narrow bandwidths that are narrower than the bandwidths used to simply separate red, green and blue wavelengths and may have pass bands of and provide a spectral resolution of less than 100 nm per band, e.g., 50-100 nm per band, and could include infrared bands from about 700 nm to about 1000 nm and/or UV bands from about 200 nm to about 400 nm. The filters 205 may also have a bandwidth resolution that is as narrow as less than 10 nm per band, for example, 5-10 nm per band.
  • Referring to FIG. 2B, image sensor 200 includes a substrate 201 on which a plurality of pixel arrays 203 and associated support circuitry (not shown) for each array 203 are fabricated. The illustrated portion shows one pixel array 203 for one color that has a corresponding color filter that only allows a limited narrow spectrum of wavelengths to pass through. Each imaging lens 204 or pixel array 203 then captures different narrow band hyperspectral information. For simplicity, the embodiment shown in FIG. 2B shows pixel arrays 203 that are four pixels wide, but it should be appreciated that each of the pixel arrays 203 may contain hundreds, thousands or millions of pixels, as desired. Opaque walls 260 separate the individual arrays 203 and opaque walls 270 separate the arrays of imaging lenses 204 that are arranged above each respective pixel array 203. Opaque walls may also optionally separate the narrow hyperspectral filters 205. The hyperspectral bandpass filters 205 are unique for each imaging lens 204, meaning that each imaging lens 204 or pixel array 203 under each lens 204 has a different narrow band color filter that only allows a limited spectrum of wavelengths to pass through. Thus, each pixel array 203 captures different narrow band hyperspectral information.
  • As noted, the pixel arrays 203 preferably are integrated on a single silicon die substrate with common circuitry. Including a multi-array color image sensor on a single die provides for a reduction of color crosstalk artifacts, especially for compact camera modules with pixel sizes less than 6 microns by 6 microns. Moreover, an imaging lens with a short focal length can minimize parallax effects and allow a camera module employing image sensor 200 to be more compact.
  • Although FIG. 2B illustrates a single lens 204 over each array 203, a modified embodiment provides an imaging lens 204 over one or more pixels in each array 203. For example, each individual lens 204 may cover and focus light on a sixteen pixel section (in a 4×4 pattern) of the pixels in each array 203.
  • Because the image sensor 200 employs hyperspectral imaging, the image data is simultaneously collected in dozens or hundreds of narrow, adjacent hyperspectral bands, as illustrated in FIG. 3. For example, a 4×4 arrangement of pixel arrays (as illustrated in FIGS. 2A-2B) has at most 16 hyperspectral bands. For collection of hundreds of hyperspectral bands, the image sensor must have a hundred or more individual pixel arrays, e.g., a 10×10 arrangement of pixel arrays.
  • Each image I1 . . . In from a respective array represents a hyperspectral image from a respective pixel array 203 (FIG. 2A). The totality of images I1 . . . In represent an image with all hyperspectral bands, as illustrated in FIG. 4. The individual images can then be combined and thus, each pixel in the combined image would contain full hyperspectral information. In FIG. 3, β1 . . . βn represents the bandpass optical filter for each image I1 . . . In of FIG. 4. Each imaging lens 204 with its unique hyperspectral bandpass filter will image the same scene if the image sensor arrays 203 are in close proximity to one another, e.g., on the order of less than 2 mm and with a preferred spacing in the range of about 0.5 mm to about 1 mm. For arrays 203 which have respective lenses 204, the lenses are similarly spaced. Thus, a hyperspectral or full spectral image may be constructed such that each pixel in the produced image contains a full spectral representation of reflectance of that point in the scene. Further, because the image sensor arrays 203 and imaging lenses 204 are in close proximity, e.g., on the order of less than 2 mm and preferably about 0.5 mm to about 1 mm, then object parallax is quite small.
  • FIG. 5 is a simplified cross sectional view of a portion of an image sensor 200 according to any embodiment of the invention. Sensor 200 includes a substrate 201 on which a plurality of pixel arrays 203 and associated support circuitry (not shown) for each array are fabricated. Each color pixel array 203 is associated with a different color on a wavelength spectrum. As discussed above, each imaging lens 204 has a unique hyperspectral bandpass filter 205 that is associated with each pixel array 203 which is designed to pass a limited spectrum of wavelengths. Each pixel array 203 then captures different hyperspectral information.
  • As an object to be imaged moves closer to the array of imaging lenses 204, the individual arrays 203 will exhibit an increase in parallax distance between them. The magnitude of the parallax distance between two adjacent arrays is approximated by the following formula:

  • P x=(N/x)*d x   (1)
  • where N is the pixel array dimension (in pixels), dx is the distance between the outer focal points of lenses 204 and where x is approximated by the following formula:

  • x=2*O*Tan(α/2)   (2)
  • where O is the object distance from the camera (or distance from the object to be imaged and the imaging lenses 204) and α is the field of view (FOV) angle. Once the object distance O has been measured or approximated, the parallax distance calculation can be performed. An example for calculating parallax is discussed below. This example also shows that small pixel arrays (on the order of about 1 mm and spaced less than 2 mm apart, e.g., from about 0.5 mm to about 1 mm) will not produce excessive parallax.
  • For an object distance O of about 2 m from a camera and a field of view angle of 60° , the Px or parallax distance at the object plane between adjacent pixel arrays 203 would equal 0.78 pixels by using formulas (1) and (2) above. Thus, for image sensors with 4×4 imaging arrays (i.e., 4×4 image sensor), the maximum parallax (at 60° FOV) between images for objects with a distance of greater than or equal to 2 m from the camera would equal 2.35 pixels. A 4×4 image sensor would have 16 hyperspectral bands, a pixel element size of 2.2 μm×2.2 μm and a target hyperspectral image size of 800×600.
  • Various image sensor configurations allow tradeoffs in performance. For example, a 3×3 image sensor will have less maximum parallax than a 4×4 image sensor, but will have only 9 hyperspectral bands. A larger FOV will have less parallax but the user must get closer to objects they want to isolate in a scene. Further, smaller pixels may be used to reduce maximum parallax with a tradeoff of a reduction in camera sensitivity.
  • FIG. 6 illustrates a block diagram of a CMOS imaging device 500, which employs a multi-array sensor 200 having a plurality of pixel arrays 203 constructed according to any embodiment described above. Multiple arrays 203 are arranged to form one large pixel array 505. The pixels of each row in array 505 are all turned on at the same time by a row select line, and the pixels of each column are selectively output onto the output lines by respective column select lines. A plurality of row and column select lines are provided for the entire array 505. The row lines are selectively activated in sequence by a row driver 510 in response to row address decoder 515. The column select lines are selectively activated in sequence for each row activation by a column driver 520 in response to column address decoder 525. Thus, a row and column address is provided for each pixel.
  • The CMOS image sensor 500 is operated by a control circuit 530, which controls address decoders 515, 525 for selecting the appropriate row and column select lines for pixel readout. Control circuit 530 also controls the row and column driver circuitry 510, 520 so that they apply driving voltages to the drive transistors of the selected row and column select lines. The pixel output signals typically include a pixel reset signal Vrst read out of the storage region after it is reset by the reset transistor and a pixel image signal Vsig, which is read out of the storage region after photo-generated charges are transferred to the region. The Vrst and Vsig signals are sampled by a sample and hold circuit 535 and are subtracted by a differential amplifier 540, to produce a differential signal Vrst-Vsig for each readout pixel. Vrst-Vsig represents the amount of light impinging on the pixels. This difference signal is digitized by an analog-to-digital converter 545. The digitized pixel signals are fed to an image processor 550 to form a digital image output. The digitizing and image processing can be located on or off the imaging device chip. In some arrangements the differential signal Vrst-Vsig can be amplified as a differential signal and directly digitized by a differential analog to digital converter. The analog-to digital converter 545 supplies the digitized pixel signals to an image processor 550, which performs appropriate image processing, which can include combining the outputs of multiple arrays and performing a parallax adjustment calculation if needed or desired, before providing digital signals defining an image output.
  • It should be noted that FIG. 6 represents one example of a readout circuit for multi-array 505. Another embodiment could employ the readout circuit in FIG. 6 for each individual array 203 with an image processor 550 combining the outputs of the individual arrays 203. Also, FIG. 6 illustrates a readout circuit suitable for CMOS arrays 203, but the invention may be used with other solid state imaging arrays, for example, CCD arrays in which case a readout circuit suitable for reading out CCD arrays could be employed.
  • FIG. 7 illustrates a processor system 900 including the image sensor 500 of FIG. 6. The processor system 900 is one example of a system having digital circuits that could include image sensor devices. Without being limiting, such a processor system could also include a computer system, camera system, scanner, machine vision, vehicle navigation, video phone, surveillance system, auto focus system, star tracker system, motion detection system, image stabilization system, and other processor system.
  • The processor system 900, for example a digital camera system, generally comprises a central processing unit (CPU) 995, such as a microprocessor for common operational control, that communicates with an input/output (I/O) device 991 over a bus 993. Image sensor 500 also communicates with the CPU 995 over the bus 993. The processor-based system 900 also includes random access memory (RAM) 992, and can include removable memory 994, such as flash memory, which also communicate with CPU 995 over the bus 993. Image sensor 500 may be combined with a processor, such as a CPU, digital signal processor, or microprocessor, with or without memory storage on a single integrated circuit or on a different chip than the processor. A parallax adjustment calculation may be performed by the image processor 550 in image sensor 500, or by the CPU 995.
  • While the embodiments have been described in detail in connection with the embodiments known at the time, it should be readily understood that the claimed invention is not limited to the disclosed embodiments. Rather, they can be modified to incorporate any number of variations, alterations, substitutions or equivalent arrangements not heretofore described. For example, while embodiments are described in connection with a CMOS pixel image sensor, they can be practiced with any other type of pixel image sensor (e.g., CCD, etc.). Furthermore, the various embodiments could be used in automotive applications and other applications where the object plane is at a constant distance, parallax is easily accounted for by using a simple linear shift of pixel data from each camera to properly register all images, such as in machine vision or industrial imaging. In addition, various embodiments may be used in low-cost, solid-state hyperspectral scanners, in which multiple hyperspectral cameras in one scanning system may produce high-resolution images very quickly.

Claims (48)

1. An imager apparatus comprising:
a plurality of pixel arrays on a single die; and
a plurality of imaging lenses for focusing an image on the pixel arrays, each of the pixel arrays being associated with a respective hyperspectral bandpass filter.
2. The imager apparatus of claim 1, wherein each of the hyperspectral filters has a spectral bandpass of less than 100 nm.
3. The imager apparatus of claim 2, wherein each of the hyperspectral filters has a spectral bandpass of between 50 nm and 100 nm.
4. The imager apparatus of claim 2, wherein each of the hyperspectral filters has a spectral bandpass of less than 10 nm.
5. The imager apparatus of claim 4, wherein each of the hyperspectral filters has a spectral bandpass of between 5 nm and 10 nm.
6. The imager apparatus of claim 1, wherein each hyperspectral bandpass filter is integrated with its associated imaging lens.
7. The imager apparatus of claim 6, wherein each hyperspectral bandpass filter is a thin film coating on its associated lens.
8. The imager apparatus of claim 1, wherein each hyperspectral bandpass filter is a separate element from its associated imaging lens.
9. The imager apparatus of claim 1, wherein each hyperspectral bandpass filter is a thin film coating on a surface of its associated pixel array.
10. The imager apparatus of claim 1, wherein each hyperspectral bandpass filter has a unique bandpass range.
11. The imager apparatus of claim 1, wherein the hyperspectral bandpass filters collectively cover a spectrum of between about 700 nm and about 1000 nm.
12. The imager apparatus of claim 1, wherein the hyperspectral bandpass filters collectively cover a spectrum of between about 200 nm and about 400 nm.
13. The imager apparatus of claim 1, wherein the plurality of pixel arrays comprise M×N pixel arrays arranged in a M×N pattern of arrays on the die.
14. The imager apparatus of claim 13, wherein M equals N.
15. The imager apparatus of claim 13, wherein M does not equal N.
16. The imager apparatus of claim 1, wherein the pixel arrays are spaced less than 2 mm apart.
17. The imager apparatus of claim 16, wherein the pixel arrays are spaced about 0.5 mm to about 1 mm apart.
18. The imager apparatus of claim 16, wherein the imaging lenses and associated hyperspectral bandpass filters are configured to capture a full image of a scene.
19. The imager apparatus of claim 1, further comprising an optical barrier between adjacent pixel arrays.
20. The imager apparatus of claim 1, further comprising an optical barrier between adjacent imaging lenses.
21. An imager apparatus comprising:
a plurality of pixel arrays formed on a single die, wherein the plurality of arrays are spaced apart by less than 2 millimeters;
a plurality of hyperspectral bandpass filters respectively associated with the pixel arrays, each of the hyperspectral filters having a spectral bandpass of less than 100 nm; and
a plurality of imaging lenses respectively associated with the pixel arrays.
22. The imager apparatus of claim 21, wherein each of the hyperspectral filters has a spectral bandpass of between 50 nm and 100 nm.
23. The imager apparatus of claim 21, wherein each of the hyperspectral filters has a spectral bandpass of less than 10 nm.
24. The imager apparatus of claim 23, wherein each of the hyperspectral filters has a spectral bandpass of between 5 nm and 10 nm.
25. The imager apparatus of claim 21, wherein each imaging lens has a hyperspectral bandpass filter associated with its respective imaging lens.
26. The imager apparatus of claim 25, wherein each hyperspectral bandpass filter is integrated with its associated imaging lens.
27. The imager apparatus of claim 26, wherein each hyperspectral bandpass filter is a thin film coating on its associated lens.
28. The imager apparatus of claim 25, wherein each hyperspectral bandpass filter is a separate element from its associated imaging lens.
29. The imager apparatus of claim 21, wherein each hyperspectral bandpass filter is a thin film coating on a surface of its associated pixel array.
30. The imager apparatus of claim 21, wherein each hyperspectral bandpass filter has a unique bandpass range.
31. The imager apparatus of claim 21, wherein the plurality of pixel arrays comprise M×N pixel arrays arranged in a M×N pattern of arrays on the die.
32. The imager apparatus of claim 31, wherein M equals N.
33. The imager apparatus of claim 31, wherein M does not equal N.
34. The imager apparatus of claim 21, wherein the pixel arrays and the imaging lenses are spaced about 0.5 mm to about 1 mm apart.
35. The imager apparatus of claim 34, wherein the imaging lenses and associated hyperspectral bandpass filters are configured to capture a full image of a scene.
36. An imager apparatus comprising:
an image sensor comprising:
a plurality of pixel arrays;
a plurality of imaging lenses respectively arranged above the pixel arrays; and
a hyperspectral bandpass filter associated with each imaging lens, wherein each hyperspectral bandpass filter is unique for each imaging lens; and
a readout circuit for reading out respective image signals from each of the arrays.
37. The imager apparatus of claim 36, further comprising:
an image combining device for combining the respective image signals from each of the pixel arrays.
38. The imager apparatus of claim 37, wherein the image combining device performs a parallax adjustment calculation.
39. The imager apparatus of claim 37, wherein a common readout circuit is provided for all of the pixel arrays.
40. The imager apparatus of claim 37, wherein separate readout circuits are provided for each of the pixel arrays.
41. The imager apparatus of claim 36, wherein the plurality of pixel arrays are formed on a single die.
42. The imager apparatus of claim 36, wherein each hyperspectral bandpass filter is integrated with its associated imaging lens.
43. The imager apparatus of claim 42, wherein each hyperspectral bandpass filter is a thin film coating on its associated lens.
44. The imager apparatus of claim 36, wherein each hyperspectral bandpass filter is a separate element from its associated imaging lens.
45. The imager apparatus of claim 36, wherein each hyperspectral bandpass filter is a thin film coating on a surface of its associated pixel array.
46. The imager apparatus of claim 36, wherein the pixel arrays are spaced less than 2 mm apart.
47. The imager apparatus of claim 46, wherein the pixel arrays are spaced about 0.5 mm to about 1 mm apart.
48. The imager apparatus of claim 36, wherein the imager apparatus is a component of a digital camera.
US11/642,867 2006-03-06 2006-12-21 Method, apparatus and system providing an integrated hyperspectral imager Abandoned US20070206242A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/642,867 US20070206242A1 (en) 2006-03-06 2006-12-21 Method, apparatus and system providing an integrated hyperspectral imager

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US11/367,580 US7924483B2 (en) 2006-03-06 2006-03-06 Fused multi-array color image sensor
US11/642,867 US20070206242A1 (en) 2006-03-06 2006-12-21 Method, apparatus and system providing an integrated hyperspectral imager

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/367,580 Continuation-In-Part US7924483B2 (en) 2006-03-06 2006-03-06 Fused multi-array color image sensor

Publications (1)

Publication Number Publication Date
US20070206242A1 true US20070206242A1 (en) 2007-09-06

Family

ID=38193548

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/367,580 Active 2030-02-10 US7924483B2 (en) 2006-03-06 2006-03-06 Fused multi-array color image sensor
US11/642,867 Abandoned US20070206242A1 (en) 2006-03-06 2006-12-21 Method, apparatus and system providing an integrated hyperspectral imager

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/367,580 Active 2030-02-10 US7924483B2 (en) 2006-03-06 2006-03-06 Fused multi-array color image sensor

Country Status (7)

Country Link
US (2) US7924483B2 (en)
EP (1) EP2008470A1 (en)
JP (1) JP2009529291A (en)
KR (1) KR20080109794A (en)
CN (1) CN101395926A (en)
TW (1) TWI398948B (en)
WO (1) WO2007103212A1 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070201114A1 (en) * 2006-02-28 2007-08-30 Yoshitaka Egawa Solid-state image sensing device having photoelectric conversion cells each configured by n pixels in vertical direction
WO2009085305A1 (en) 2007-12-27 2009-07-09 Google Inc. High-resolution, variable depth of field image device
US20120147233A1 (en) * 2010-12-09 2012-06-14 Sony Corporation Imaging device and imaging apparatus
US20140267878A1 (en) * 2011-11-04 2014-09-18 Imec Spectral camera with overlapping segments of image copies interleaved onto sensor array
US20140347533A1 (en) * 2013-05-21 2014-11-27 Olympus Corporation Image processing device and image processing method
US20150077524A1 (en) * 2012-03-16 2015-03-19 Nikon Corporation Image sensor and imaging device
US20150177429A1 (en) * 2012-06-05 2015-06-25 Hypermed Imaging, Inc. Single-sensor hyperspectral imaging device
US20150325613A1 (en) * 2013-07-30 2015-11-12 Heptagon Micro Optics Pte. Ltd. Optoelectronic modules that have shielding to reduce light leakage or stray light, and fabrication methods for such modules
US9217671B2 (en) 2009-09-01 2015-12-22 Koninklijke Philips N.V. High spectral resolution color sensor using non-dispersive elements
US9772229B2 (en) 2011-11-04 2017-09-26 Imec Spectral camera with integrated filters and multiple adjacent image copies projected onto sensor array
US20190190237A1 (en) * 2016-09-29 2019-06-20 Intel Corporation Optical free air bus interconnect
CN110235252A (en) * 2016-12-07 2019-09-13 兴业法兰西红外探测器-索弗拉迪尔公司 For improving the device of multispectral detection
US20200329205A1 (en) * 2020-04-02 2020-10-15 Sichuan Dualix Spectral Lmaging Technology Co., Ltd Hyperspectral camera based on continuously variable film filter and coating method thereof
US20220214217A1 (en) * 2021-01-04 2022-07-07 Argo AI, LLC Systems and methods for characterizing spectral reflectance of real world objects

Families Citing this family (124)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7955764B2 (en) * 2006-04-07 2011-06-07 Micron Technology, Inc. Methods to make sidewall light shields for color filter array
US7718533B2 (en) * 2007-05-08 2010-05-18 Micron Technology, Inc. Inverted variable resistance memory cell and method of making the same
US20090051790A1 (en) * 2007-08-21 2009-02-26 Micron Technology, Inc. De-parallax methods and apparatuses for lateral sensor arrays
CN101459853B (en) * 2007-12-10 2011-08-10 鸿富锦精密工业(深圳)有限公司 Color processing apparatus and method
JP5173536B2 (en) * 2008-04-02 2013-04-03 シャープ株式会社 Imaging apparatus and optical axis control method
EP2283644A4 (en) * 2008-05-09 2011-10-26 Ecole Polytech Image sensor having nonlinear response
CN102037717B (en) 2008-05-20 2013-11-06 派力肯成像公司 Capturing and processing of images using monolithic camera array with hetergeneous imagers
US11792538B2 (en) 2008-05-20 2023-10-17 Adeia Imaging Llc Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US8866920B2 (en) 2008-05-20 2014-10-21 Pelican Imaging Corporation Capturing and processing of images using monolithic camera array with heterogeneous imagers
JP4952768B2 (en) * 2009-10-27 2012-06-13 株式会社ニコン Imaging apparatus and image analysis computer program
EP2502115A4 (en) 2009-11-20 2013-11-06 Pelican Imaging Corp Capturing and processing of images using monolithic camera array with heterogeneous imagers
US8319855B2 (en) * 2010-01-19 2012-11-27 Rambus Inc. Method, apparatus and system for image acquisition and conversion
KR101824672B1 (en) 2010-05-12 2018-02-05 포토네이션 케이맨 리미티드 Architectures for imager arrays and array cameras
WO2012051751A1 (en) * 2010-10-19 2012-04-26 Ether Precision, Inc. Optical module comprising monochromatic image sensors, system comprising optical module and method of manufacturing optical module
US8878950B2 (en) 2010-12-14 2014-11-04 Pelican Imaging Corporation Systems and methods for synthesizing high resolution images using super-resolution processes
US20120274811A1 (en) * 2011-04-28 2012-11-01 Dmitry Bakin Imaging devices having arrays of image sensors and precision offset lenses
JP2014519741A (en) 2011-05-11 2014-08-14 ペリカン イメージング コーポレイション System and method for transmitting and receiving array camera image data
JP2012253727A (en) 2011-06-07 2012-12-20 Toshiba Corp Solid-state image sensor and camera module
US20130265459A1 (en) 2011-06-28 2013-10-10 Pelican Imaging Corporation Optical arrangements for use with an array camera
WO2013003276A1 (en) 2011-06-28 2013-01-03 Pelican Imaging Corporation Optical arrangements for use with an array camera
WO2013043751A1 (en) 2011-09-19 2013-03-28 Pelican Imaging Corporation Systems and methods for controlling aliasing in images captured by an array camera for use in super resolution processing using pixel apertures
KR102002165B1 (en) 2011-09-28 2019-07-25 포토내이션 리미티드 Systems and methods for encoding and decoding light field image files
EP2817955B1 (en) 2012-02-21 2018-04-11 FotoNation Cayman Limited Systems and methods for the manipulation of captured light field image data
US9210392B2 (en) 2012-05-01 2015-12-08 Pelican Imaging Coporation Camera modules patterned with pi filter groups
KR20150023907A (en) 2012-06-28 2015-03-05 펠리칸 이매징 코포레이션 Systems and methods for detecting defective camera arrays, optic arrays, and sensors
US20140002674A1 (en) 2012-06-30 2014-01-02 Pelican Imaging Corporation Systems and Methods for Manufacturing Camera Modules Using Active Alignment of Lens Stack Arrays and Sensors
EP2888720B1 (en) 2012-08-21 2021-03-17 FotoNation Limited System and method for depth estimation from images captured using array cameras
WO2014032020A2 (en) 2012-08-23 2014-02-27 Pelican Imaging Corporation Feature based high resolution motion estimation from low resolution images captured using an array source
US9214013B2 (en) 2012-09-14 2015-12-15 Pelican Imaging Corporation Systems and methods for correcting user identified artifacts in light field images
US20140092281A1 (en) 2012-09-28 2014-04-03 Pelican Imaging Corporation Generating Images from Light Fields Utilizing Virtual Viewpoints
WO2014078443A1 (en) 2012-11-13 2014-05-22 Pelican Imaging Corporation Systems and methods for array camera focal plane control
US9363425B2 (en) 2012-12-06 2016-06-07 Semiconductor Components Industries, Llc Color filter arrangements for fused array imaging systems
WO2014130849A1 (en) 2013-02-21 2014-08-28 Pelican Imaging Corporation Generating compressed light field representation data
WO2014133974A1 (en) 2013-02-24 2014-09-04 Pelican Imaging Corporation Thin form computational and modular array cameras
WO2014133481A1 (en) * 2013-02-26 2014-09-04 Hewlett-Packard Development Company, L.P. Multiview 3d telepresence
US9774789B2 (en) 2013-03-08 2017-09-26 Fotonation Cayman Limited Systems and methods for high dynamic range imaging using array cameras
US8866912B2 (en) 2013-03-10 2014-10-21 Pelican Imaging Corporation System and methods for calibration of an array camera using a single captured image
US9521416B1 (en) 2013-03-11 2016-12-13 Kip Peli P1 Lp Systems and methods for image data compression
US9106784B2 (en) 2013-03-13 2015-08-11 Pelican Imaging Corporation Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing
US9888194B2 (en) 2013-03-13 2018-02-06 Fotonation Cayman Limited Array camera architecture implementing quantum film image sensors
WO2014165244A1 (en) 2013-03-13 2014-10-09 Pelican Imaging Corporation Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies
WO2014164550A2 (en) 2013-03-13 2014-10-09 Pelican Imaging Corporation System and methods for calibration of an array camera
US9100586B2 (en) 2013-03-14 2015-08-04 Pelican Imaging Corporation Systems and methods for photometric normalization in array cameras
WO2014159779A1 (en) 2013-03-14 2014-10-02 Pelican Imaging Corporation Systems and methods for reducing motion blur in images or video in ultra low light with array cameras
CN104051484B (en) * 2013-03-15 2017-09-19 豪威科技股份有限公司 Possess the imaging sensor of the pixel with increased optical crosstalk and its use
US10122993B2 (en) 2013-03-15 2018-11-06 Fotonation Limited Autofocus system for a conventional camera that uses depth information from an array camera
US9497429B2 (en) 2013-03-15 2016-11-15 Pelican Imaging Corporation Extended color processing on pelican array cameras
US9215430B2 (en) 2013-03-15 2015-12-15 Omnivision Technologies, Inc. Image sensor with pixels having increased optical crosstalk
EP2973476A4 (en) 2013-03-15 2017-01-18 Pelican Imaging Corporation Systems and methods for stereo imaging with camera arrays
US9445003B1 (en) 2013-03-15 2016-09-13 Pelican Imaging Corporation Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information
WO2014150856A1 (en) 2013-03-15 2014-09-25 Pelican Imaging Corporation Array camera implementing quantum dot color filters
US9633442B2 (en) 2013-03-15 2017-04-25 Fotonation Cayman Limited Array cameras including an array camera module augmented with a separate camera
US20150022655A1 (en) * 2013-07-19 2015-01-22 Forrest R. Ruhge Apparatus and method using a linear array of optical sensors for imaging a rotating component of a gas turbine engine
US9898856B2 (en) 2013-09-27 2018-02-20 Fotonation Cayman Limited Systems and methods for depth-assisted perspective distortion correction
US9264592B2 (en) 2013-11-07 2016-02-16 Pelican Imaging Corporation Array camera modules incorporating independently aligned lens stacks
WO2015074078A1 (en) 2013-11-18 2015-05-21 Pelican Imaging Corporation Estimating depth from projected texture using camera arrays
WO2015081279A1 (en) 2013-11-26 2015-06-04 Pelican Imaging Corporation Array camera configurations incorporating multiple constituent array cameras
WO2015134996A1 (en) 2014-03-07 2015-09-11 Pelican Imaging Corporation System and methods for depth regularization and semiautomatic interactive matting using rgb-d images
US9521319B2 (en) 2014-06-18 2016-12-13 Pelican Imaging Corporation Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor
US10250871B2 (en) 2014-09-29 2019-04-02 Fotonation Limited Systems and methods for dynamic calibration of array cameras
US9942474B2 (en) 2015-04-17 2018-04-10 Fotonation Cayman Limited Systems and methods for performing high speed video capture and depth estimation using array cameras
CN105611123B (en) * 2015-12-18 2017-05-24 广东欧珀移动通信有限公司 imaging method, image sensor, imaging device and electronic device
CN105578076A (en) * 2015-12-18 2016-05-11 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic device
CN105592303B (en) * 2015-12-18 2018-09-11 广东欧珀移动通信有限公司 imaging method, imaging device and electronic device
CN105635702B (en) * 2015-12-18 2017-06-13 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic installation
CN105578072A (en) * 2015-12-18 2016-05-11 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic device
CN105578006B (en) * 2015-12-18 2018-02-13 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic installation
CN105611125B (en) * 2015-12-18 2018-04-10 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic installation
CN105611257B (en) * 2015-12-18 2018-09-11 广东欧珀移动通信有限公司 Imaging method, imaging sensor, imaging device and electronic device
CN105578080B (en) * 2015-12-18 2019-02-05 Oppo广东移动通信有限公司 Imaging method, imaging device and electronic device
CN105430361B (en) * 2015-12-18 2018-03-20 广东欧珀移动通信有限公司 Imaging method, imaging sensor, imaging device and electronic installation
CN105578077A (en) * 2015-12-18 2016-05-11 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic device
CN105635532B (en) * 2015-12-18 2017-05-24 广东欧珀移动通信有限公司 imaging method, image sensor, imaging device and electronic device
US10683034B2 (en) * 2017-06-06 2020-06-16 Ford Global Technologies, Llc Vehicle remote parking systems and methods
US10585430B2 (en) 2017-06-16 2020-03-10 Ford Global Technologies, Llc Remote park-assist authentication for vehicles
US10775781B2 (en) 2017-06-16 2020-09-15 Ford Global Technologies, Llc Interface verification for vehicle remote park-assist
US10482618B2 (en) 2017-08-21 2019-11-19 Fotonation Limited Systems and methods for hybrid depth regularization
JP6963295B2 (en) * 2017-09-01 2021-11-05 学校法人東京電機大学 3D information acquisition device
US10580304B2 (en) 2017-10-02 2020-03-03 Ford Global Technologies, Llc Accelerometer-based external sound monitoring for voice controlled autonomous parking
US10627811B2 (en) 2017-11-07 2020-04-21 Ford Global Technologies, Llc Audio alerts for remote park-assist tethering
CN107846537B (en) 2017-11-08 2019-11-26 维沃移动通信有限公司 A kind of CCD camera assembly, image acquiring method and mobile terminal
US10578676B2 (en) 2017-11-28 2020-03-03 Ford Global Technologies, Llc Vehicle monitoring of mobile device state-of-charge
US10583830B2 (en) 2018-01-02 2020-03-10 Ford Global Technologies, Llc Mobile device tethering for a remote parking assist system of a vehicle
US10688918B2 (en) 2018-01-02 2020-06-23 Ford Global Technologies, Llc Mobile device tethering for a remote parking assist system of a vehicle
US10814864B2 (en) 2018-01-02 2020-10-27 Ford Global Technologies, Llc Mobile device tethering for a remote parking assist system of a vehicle
US11148661B2 (en) 2018-01-02 2021-10-19 Ford Global Technologies, Llc Mobile device tethering for a remote parking assist system of a vehicle
US10585431B2 (en) 2018-01-02 2020-03-10 Ford Global Technologies, Llc Mobile device tethering for a remote parking assist system of a vehicle
US10737690B2 (en) 2018-01-02 2020-08-11 Ford Global Technologies, Llc Mobile device tethering for a remote parking assist system of a vehicle
US10974717B2 (en) 2018-01-02 2021-04-13 Ford Global Technologies, I.LC Mobile device tethering for a remote parking assist system of a vehicle
US10684773B2 (en) 2018-01-03 2020-06-16 Ford Global Technologies, Llc Mobile device interface for trailer backup-assist
US10747218B2 (en) 2018-01-12 2020-08-18 Ford Global Technologies, Llc Mobile device tethering for remote parking assist
US10917748B2 (en) 2018-01-25 2021-02-09 Ford Global Technologies, Llc Mobile device tethering for vehicle systems based on variable time-of-flight and dead reckoning
US10684627B2 (en) 2018-02-06 2020-06-16 Ford Global Technologies, Llc Accelerometer-based external sound monitoring for position aware autonomous parking
US11188070B2 (en) 2018-02-19 2021-11-30 Ford Global Technologies, Llc Mitigating key fob unavailability for remote parking assist systems
US10507868B2 (en) 2018-02-22 2019-12-17 Ford Global Technologies, Llc Tire pressure monitoring for vehicle park-assist
US10732622B2 (en) 2018-04-05 2020-08-04 Ford Global Technologies, Llc Advanced user interaction features for remote park assist
US10793144B2 (en) 2018-04-09 2020-10-06 Ford Global Technologies, Llc Vehicle remote park-assist communication counters
US10493981B2 (en) 2018-04-09 2019-12-03 Ford Global Technologies, Llc Input signal management for vehicle park-assist
US10759417B2 (en) 2018-04-09 2020-09-01 Ford Global Technologies, Llc Input signal management for vehicle park-assist
US10683004B2 (en) 2018-04-09 2020-06-16 Ford Global Technologies, Llc Input signal management for vehicle park-assist
US10384605B1 (en) 2018-09-04 2019-08-20 Ford Global Technologies, Llc Methods and apparatus to facilitate pedestrian detection during remote-controlled maneuvers
US10717432B2 (en) 2018-09-13 2020-07-21 Ford Global Technologies, Llc Park-assist based on vehicle door open positions
US10821972B2 (en) 2018-09-13 2020-11-03 Ford Global Technologies, Llc Vehicle remote parking assist systems and methods
US10967851B2 (en) 2018-09-24 2021-04-06 Ford Global Technologies, Llc Vehicle system and method for setting variable virtual boundary
US10529233B1 (en) 2018-09-24 2020-01-07 Ford Global Technologies Llc Vehicle and method for detecting a parking space via a drone
US10908603B2 (en) 2018-10-08 2021-02-02 Ford Global Technologies, Llc Methods and apparatus to facilitate remote-controlled maneuvers
US10628687B1 (en) 2018-10-12 2020-04-21 Ford Global Technologies, Llc Parking spot identification for vehicle park-assist
US11097723B2 (en) 2018-10-17 2021-08-24 Ford Global Technologies, Llc User interfaces for vehicle remote park assist
US11137754B2 (en) 2018-10-24 2021-10-05 Ford Global Technologies, Llc Intermittent delay mitigation for remote vehicle operation
CN109579752B (en) * 2018-11-20 2021-07-23 维沃移动通信有限公司 Measuring method and terminal equipment
US11789442B2 (en) 2019-02-07 2023-10-17 Ford Global Technologies, Llc Anomalous input detection
US11195344B2 (en) 2019-03-15 2021-12-07 Ford Global Technologies, Llc High phone BLE or CPU burden detection and notification
US11275368B2 (en) 2019-04-01 2022-03-15 Ford Global Technologies, Llc Key fobs for vehicle remote park-assist
US11169517B2 (en) 2019-04-01 2021-11-09 Ford Global Technologies, Llc Initiation of vehicle remote park-assist with key fob
KR102646521B1 (en) 2019-09-17 2024-03-21 인트린식 이노베이션 엘엘씨 Surface modeling system and method using polarization cue
CN114766003B (en) 2019-10-07 2024-03-26 波士顿偏振测定公司 Systems and methods for enhancing sensor systems and imaging systems with polarization
CN114787648B (en) 2019-11-30 2023-11-10 波士顿偏振测定公司 Systems and methods for transparent object segmentation using polarization cues
US11470287B2 (en) 2019-12-05 2022-10-11 Samsung Electronics Co., Ltd. Color imaging apparatus using monochrome sensors for mobile devices
KR20210081767A (en) * 2019-12-24 2021-07-02 삼성전자주식회사 Imaging device and image sensing method
CN115552486A (en) 2020-01-29 2022-12-30 因思创新有限责任公司 System and method for characterizing an object pose detection and measurement system
WO2021154459A1 (en) 2020-01-30 2021-08-05 Boston Polarimetrics, Inc. Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images
JP2023539884A (en) * 2020-08-31 2023-09-20 トライアイ リミテッド Shortwave infrared focal plane array, and methods for its use and manufacturing
US11290658B1 (en) 2021-04-15 2022-03-29 Boston Polarimetrics, Inc. Systems and methods for camera exposure control
US11689813B2 (en) 2021-07-01 2023-06-27 Intrinsic Innovation Llc Systems and methods for high dynamic range imaging using crossed polarizers

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5570221A (en) * 1993-12-28 1996-10-29 Nec Corporation Light amplification device
US5729011A (en) * 1995-02-24 1998-03-17 Olympus Optical Co., Ltd. Spectroscopic apparatus and spectroscopic image recording apparatus
US20020090650A1 (en) * 2000-04-06 2002-07-11 Quantum Dot Corporation Two-dimensional spectral imaging system
US6639664B2 (en) * 1998-05-13 2003-10-28 Storz Endoskop Gmbh Endoscope for inspection of an observation cavity
US20050057491A1 (en) * 2003-08-28 2005-03-17 Eastman Kodak Company Private display system
US6937279B1 (en) * 1998-12-22 2005-08-30 Hynix Semiconductor Inc. Apparatus for converting analog image data into digital image data in CMOS image sensor
US6961080B2 (en) * 1997-04-09 2005-11-01 Richardson Technologies Inc. Color translating UV microscope
US20070034777A1 (en) * 2005-08-12 2007-02-15 Tessera, Inc. Image sensor employing a plurality of photodetector arrays and/or rear-illuminated architecture
US7242478B1 (en) * 2003-12-05 2007-07-10 Surface Optics Corporation Spatially corrected full-cubed hyperspectral imager
US20070260413A1 (en) * 2004-10-28 2007-11-08 Peter Ehbets Method for Correcting Measured Image Values
US20080204744A1 (en) * 2005-07-11 2008-08-28 Jose Mir High Speed, Optically-Multiplexed, Hyperspectral Imagers and Methods Thereof
US7605958B2 (en) * 2004-06-02 2009-10-20 Xerox Corporation Design parameters for a multi-row linear photosensor array

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6650361B1 (en) * 1997-12-17 2003-11-18 Canon Kabushiki Kaisha Imaging apparatus control method, and a computer program product having computer program code therefor
JP3451195B2 (en) * 1998-05-28 2003-09-29 富士写真フイルム株式会社 Image reading device
US6069973A (en) * 1998-06-30 2000-05-30 Xerox Corporation Method and apparatus for color correction in a multi-chip imaging array
US6667818B1 (en) * 2000-10-23 2003-12-23 Umax Data Systems, Inc. Multiple-field sensor for scanning system
JP3703424B2 (en) * 2000-10-25 2005-10-05 キヤノン株式会社 IMAGING DEVICE, ITS CONTROL METHOD, CONTROL PROGRAM, AND STORAGE MEDIUM
JP2002135796A (en) * 2000-10-25 2002-05-10 Canon Inc Imaging apparatus
JP2003143459A (en) * 2001-11-02 2003-05-16 Canon Inc Compound-eye image pickup system and device provided therewith
JP2003283907A (en) * 2002-03-20 2003-10-03 Japan Science & Technology Corp Imaging device
EP2466871A3 (en) 2003-10-22 2017-05-03 Panasonic Intellectual Property Management Co., Ltd. Imaging apparatus and method for producing the same, portable equipment, and imaging sensor and method for producing the same.
JP2005176040A (en) * 2003-12-12 2005-06-30 Canon Inc Imaging device
US7511749B2 (en) 2003-12-18 2009-03-31 Aptina Imaging Corporation Color image sensor having imaging element array forming images on respective regions of sensor elements
US7773143B2 (en) 2004-04-08 2010-08-10 Tessera North America, Inc. Thin color camera having sub-pixel resolution
US7830426B2 (en) * 2005-12-29 2010-11-09 Micron Technology, Inc. Method and apparatus providing color interpolation in color filter arrays using edge detection and correction terms

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5570221A (en) * 1993-12-28 1996-10-29 Nec Corporation Light amplification device
US5729011A (en) * 1995-02-24 1998-03-17 Olympus Optical Co., Ltd. Spectroscopic apparatus and spectroscopic image recording apparatus
US6961080B2 (en) * 1997-04-09 2005-11-01 Richardson Technologies Inc. Color translating UV microscope
US6639664B2 (en) * 1998-05-13 2003-10-28 Storz Endoskop Gmbh Endoscope for inspection of an observation cavity
US6937279B1 (en) * 1998-12-22 2005-08-30 Hynix Semiconductor Inc. Apparatus for converting analog image data into digital image data in CMOS image sensor
US20020090650A1 (en) * 2000-04-06 2002-07-11 Quantum Dot Corporation Two-dimensional spectral imaging system
US20050057491A1 (en) * 2003-08-28 2005-03-17 Eastman Kodak Company Private display system
US7242478B1 (en) * 2003-12-05 2007-07-10 Surface Optics Corporation Spatially corrected full-cubed hyperspectral imager
US7605958B2 (en) * 2004-06-02 2009-10-20 Xerox Corporation Design parameters for a multi-row linear photosensor array
US20070260413A1 (en) * 2004-10-28 2007-11-08 Peter Ehbets Method for Correcting Measured Image Values
US20080204744A1 (en) * 2005-07-11 2008-08-28 Jose Mir High Speed, Optically-Multiplexed, Hyperspectral Imagers and Methods Thereof
US20070034777A1 (en) * 2005-08-12 2007-02-15 Tessera, Inc. Image sensor employing a plurality of photodetector arrays and/or rear-illuminated architecture

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070201114A1 (en) * 2006-02-28 2007-08-30 Yoshitaka Egawa Solid-state image sensing device having photoelectric conversion cells each configured by n pixels in vertical direction
US8854420B2 (en) 2007-12-27 2014-10-07 Google Inc. High-resolution, variable depth of field image device
WO2009085305A1 (en) 2007-12-27 2009-07-09 Google Inc. High-resolution, variable depth of field image device
EP2238745A1 (en) * 2007-12-27 2010-10-13 Google, Inc. High-resolution, variable depth of field image device
US20110134224A1 (en) * 2007-12-27 2011-06-09 Google Inc. High-Resolution, Variable Depth of Field Image Device
EP2238745A4 (en) * 2007-12-27 2012-02-22 Google Inc High-resolution, variable depth of field image device
US8319822B2 (en) 2007-12-27 2012-11-27 Google Inc. High-resolution, variable depth of field image device
US9217671B2 (en) 2009-09-01 2015-12-22 Koninklijke Philips N.V. High spectral resolution color sensor using non-dispersive elements
US8558933B2 (en) * 2010-12-09 2013-10-15 Sony Corporation Imaging device and imaging apparatus
US20120147233A1 (en) * 2010-12-09 2012-06-14 Sony Corporation Imaging device and imaging apparatus
US20140267878A1 (en) * 2011-11-04 2014-09-18 Imec Spectral camera with overlapping segments of image copies interleaved onto sensor array
US9772229B2 (en) 2011-11-04 2017-09-26 Imec Spectral camera with integrated filters and multiple adjacent image copies projected onto sensor array
US9366573B2 (en) * 2011-11-04 2016-06-14 Imec Leuven Spectral camera with overlapping segments of image copies interleaved onto sensor array
US20150077524A1 (en) * 2012-03-16 2015-03-19 Nikon Corporation Image sensor and imaging device
US10018758B2 (en) * 2012-06-05 2018-07-10 Hypermed Imaging, Inc. Single-sensor hyperspectral imaging device
US11092725B2 (en) 2012-06-05 2021-08-17 Samsung Electronics Co., Ltd. Single-sensor hyperspectral imaging device
US9766382B2 (en) * 2012-06-05 2017-09-19 Hypermed Imaging, Inc. Single-sensor hyperspectral imaging device
US20150177429A1 (en) * 2012-06-05 2015-06-25 Hypermed Imaging, Inc. Single-sensor hyperspectral imaging device
US20180275325A1 (en) * 2012-06-05 2018-09-27 Hypermed Imaging, Inc. Single-sensor hyperspectral imaging device
US10534116B2 (en) * 2012-06-05 2020-01-14 Hypermed Imaging, Inc. Single-sensor hyperspectral imaging device
US11493675B2 (en) 2012-06-05 2022-11-08 Samsung Electronics Co., Ltd. Single-sensor hyperspectral imaging device
US9392237B2 (en) * 2013-05-21 2016-07-12 Olympus Corporation Image processing device and image processing method
US20140347533A1 (en) * 2013-05-21 2014-11-27 Olympus Corporation Image processing device and image processing method
US10186540B2 (en) * 2013-07-30 2019-01-22 Heptagon Micro Optics Pte. Ltd. Optoelectronic modules that have shielding to reduce light leakage or stray light
US20150325613A1 (en) * 2013-07-30 2015-11-12 Heptagon Micro Optics Pte. Ltd. Optoelectronic modules that have shielding to reduce light leakage or stray light, and fabrication methods for such modules
US20190190237A1 (en) * 2016-09-29 2019-06-20 Intel Corporation Optical free air bus interconnect
CN110235252A (en) * 2016-12-07 2019-09-13 兴业法兰西红外探测器-索弗拉迪尔公司 For improving the device of multispectral detection
US20200329205A1 (en) * 2020-04-02 2020-10-15 Sichuan Dualix Spectral Lmaging Technology Co., Ltd Hyperspectral camera based on continuously variable film filter and coating method thereof
US11622084B2 (en) * 2020-04-02 2023-04-04 Jiangsu Dualix Spectral Imaging Technology Co., Ltd Hyperspectral camera based on continuously variable film filter and coating method thereof
US20220214217A1 (en) * 2021-01-04 2022-07-07 Argo AI, LLC Systems and methods for characterizing spectral reflectance of real world objects
US11536608B2 (en) * 2021-01-04 2022-12-27 Argo AI, LLC Systems and methods for characterizing spectral reflectance of real world objects

Also Published As

Publication number Publication date
KR20080109794A (en) 2008-12-17
TW200742054A (en) 2007-11-01
US20070206241A1 (en) 2007-09-06
WO2007103212A1 (en) 2007-09-13
JP2009529291A (en) 2009-08-13
TWI398948B (en) 2013-06-11
US7924483B2 (en) 2011-04-12
EP2008470A1 (en) 2008-12-31
CN101395926A (en) 2009-03-25

Similar Documents

Publication Publication Date Title
US20070206242A1 (en) Method, apparatus and system providing an integrated hyperspectral imager
US11916093B2 (en) Solid-state imaging device, driving method therefor, and electronic apparatus
US10297629B2 (en) Image sensors with in-pixel lens arrays
US7242478B1 (en) Spatially corrected full-cubed hyperspectral imager
US7433042B1 (en) Spatially corrected full-cubed hyperspectral imager
KR101890940B1 (en) Imaging device and imaging apparatus
TWI395324B (en) Method, apparatus and system providing holographic layer as micro-lens and color filter array in an imager
US20080080028A1 (en) Imaging method, apparatus and system having extended depth of field
US9666631B2 (en) Photodiode and filter configuration for high dynamic range image sensor
US20080170228A1 (en) Method and apparatus for wafer level calibration of imaging sensors
US6661457B1 (en) Pixel read-out architecture
US20120062774A1 (en) Adaptive solid state image sensor
JP6348271B2 (en) Mixed material multispectral Stirling array sensor
KR20080091023A (en) Solid-state imaging device and imaging apparatus
US10165211B1 (en) Image sensors with optically black pixels
US10942304B2 (en) Solid-state imaging element, manufacturing method of the same, and electronic device
US9921106B1 (en) Integrated imaging spectrometer for hyperspectral imaging systems
US8416327B2 (en) Solid-state image pickup apparatus
US8077230B2 (en) Methods and apparatus for reducing color material related defects in imagers
US20050109917A1 (en) Multi-spectral imaging with almost-full fill-factor using 3D pixels
CN113824906A (en) Image sensing device
KR20220041351A (en) Image sensing device
US20210266431A1 (en) Imaging sensor pixels having built-in grating
US11817468B2 (en) Image sensing device
US11917272B2 (en) Imaging systems for multi-spectral imaging

Legal Events

Date Code Title Description
AS Assignment

Owner name: MICRON TECHNOLOGY, INC., IDAHO

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SMITH, SCOTT;REEL/FRAME:018717/0427

Effective date: 20061218

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION