WO1991007655A1 - Imaging using scattered and diffused radiation - Google Patents

Imaging using scattered and diffused radiation Download PDF

Info

Publication number
WO1991007655A1
WO1991007655A1 PCT/US1990/006664 US9006664W WO9107655A1 WO 1991007655 A1 WO1991007655 A1 WO 1991007655A1 US 9006664 W US9006664 W US 9006664W WO 9107655 A1 WO9107655 A1 WO 9107655A1
Authority
WO
WIPO (PCT)
Prior art keywords
coefficients
values
computing
attenuation
computed
Prior art date
Application number
PCT/US1990/006664
Other languages
French (fr)
Inventor
Jerome Ralph Singer
Francisco Alberto Grunbaum
Philip David Kohn
Jorge Passamani Zubelli
John Loyd Couch
Harold Lance Naparst
Geoffrey Latham
Original Assignee
Singer Imaging, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Singer Imaging, Inc. filed Critical Singer Imaging, Inc.
Priority to JP91501478A priority Critical patent/JPH05502393A/en
Publication of WO1991007655A1 publication Critical patent/WO1991007655A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/003Reconstruction from projections, e.g. tomography
    • G06T11/006Inverse problem, transformation from projection-space into object-space, e.g. transform methods, back-projection, algebraic methods
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/17Systems in which incident light is modified in accordance with the properties of the material investigated
    • G01N21/47Scattering, i.e. diffuse reflection
    • G01N21/4795Scattering, i.e. diffuse reflection spatially resolved investigating of object in scattering medium
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/17Systems in which incident light is modified in accordance with the properties of the material investigated
    • G01N2021/178Methods for obtaining spatial resolution of the property being measured
    • G01N2021/1785Three dimensional
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2211/00Image generation
    • G06T2211/40Computed tomography
    • G06T2211/424Iterative

Definitions

  • the present invention relates generally to systems which produce images of the interior of an object and particularly to methods and systems for imaging the interior of an object using scattered radiation or scattered particles.
  • the present invention uses radiation, such as photons, phonons, neutrons or other particles, which will be scattered to a significant degree by the internal structures in various objects.
  • the present invention radiates an object and measures the attenuated and scattered radiation at a number of points along the exterior of the object. The inventors have found that these measurements are sufficient to determine the scattering and attenuation properties of the various regions inside the object.
  • the interior of the object is modelled as an array of volume elements, herein called voxels.
  • Each voxel in the model of the object has scattering and attenuation properties which are represented by numerical parameters that can be mapped so as to generate several images of the interior of the object.
  • the technique of the present invention has not been tried in the past because it has been assumed, at least implicitly, that the scattered light exiting an object would simply not contain sufficient information to generate an image of the interior of that object. For instance, any one photon travelling through the object may be scattered multiple times. In addition, photons travelling on a number of different paths can emerge at the same point on the exterior of the body. Thus, it would appear that the amount of light exiting an object at an array of points along the outside of the object would not contain sufficient information to uniquely identify the scattering and attenuation properties of the object's interior. However, the inventors have made the surprising discovery that there is, in fact, more than sufficient information to determine the internal properties of the object.
  • the present invention is a system for generating an image of the interior of an object using radiation which is scattered and attenuated by the interior of the object.
  • the imaging system includes a subsystem for sequentially radiating an object with light from an array of entry points near the exterior of the object, and for measuring the intensity of light emerging from the object at exit points near the exterior of the object each time that the object is radiated with light at a different entry point.
  • the interior of the object is modelled as an array of volume elements, herein called voxels.
  • Each voxel in the model of the object has scattering and attenuation properties which are represented by numerical coefficients that can be mapped so as to generate several images of the interior of the object.
  • the scattering and attenuation coefficients for all the voxels are assigned initial values, which helps to shorten the computation process.
  • the system computes the intensity of light that would emerge from the object at the exit points if the interior of the object were characterized by the currently assigned values for the scattering and attenuation coefficients. Then, the differences between the measured light intensities and the computed light intensities are used to compute an error function related to the magnitude of the errors in the reconstruction. This "error" or "cost” function is then minimized using a gradient descent methodology, i.e., the value of the coefficients are modified so as to reduce the value of the error function.
  • the process of computing exiting light intensities based on the currently assigned values for the scattering and attenuation coefficients and then comparing the differences between the computed values and the measured values to generate a new approximation of the scattering and attenuation properties of the object continues until the error function (which corresponds to the differences between the measured and computed exiting light intensities) falls below a specified threshold value.
  • the final values of the scattering and attenuation coefficients from this process are then mapped so as to generate a series of images of the interior of the object depicting the attenuation and scattering characteristics of the object's interior.
  • FIG. 1 is a block diagram of an imaging system in accordance with the present invention.
  • Figure 2 schematically depicts how an object is modelled as an array of volume elements.
  • Figure 3 schematically depicts the attenuation, scattering and reflection properties of a single volume element.
  • Figure 4 is a block diagram of a measurement subsystem in accordance with the present invention.
  • Figure 5 is a flow chart of the imaging method used in the preferred embodiment of the invention.
  • Figure 6 depicts a map of the attenuation characteristics of an object.
  • Figure 7 depicts a map of a reconstruction of the object shown in Figure 6 generated using a prototype of the present invention.
  • Figures 8A-8C depict a map of the forward scattering characteristics of a three dimensional object and two maps of reconstructions of the object with varying amounts of noise in the measurement signal.
  • Figures 9A-9C depict a map of the attenuation characteristics of a three dimensional object and two maps of reconstructions of the object with varying amounts of noise in the measurement signal.
  • Figures 10A-10C depict a map of the side scattering characteristics of a three dimensional object and two maps of reconstructions of the object with varying amounts of noise in the measurement signal.
  • an imaging system 100 which generates an image of the internal structure of an object 102 based on measurements of scattered radiation which passes through the object.
  • a measurement subsystem 110 radiates the object 102 with radiation, such as visible or infrared light, which will be scattered to a significant degree by the internal structures in the object.
  • the measurement subsystem 110 includes a radiator system 112 (e.g., a particle radiator or a lighting system) for sequentially radiating an object with light from an array of entry points near the exterior of the object, and a sensor system 114 (e.g., an array of particle detectors or light sensors) for measuring the intensity of light emerging from the object at an array exit points near the exterior of the object each time that the object is radiated with light at a different entry point.
  • the measurement data generated by the measurement subsystem 110 is stored in a data storage device 116 and processed by a digital computer 120.
  • the computer 120 generates positioning control signals A and B for controlling the set of positions of the radiation source(s) 112 and detectors 114, respectively, so as to obtain data sets sufficient for reconstructing an image of the interior structure of the object 102.
  • a software model 122 is used to represent the interior of the object as an array of volume elements, called voxels.
  • the interior of the object 122 is modelled by a set of numerical parameters 124 which include scattering and attenuation coefficients for each of the voxels. These coefficients are assigned initial values by either the model 122, or by a user via the user interface 126.
  • the model 122 in the preferred embodiment initializes the coefficients in all the voxels with the same set of values, which corresponds to an object with a totally uniform interior. However, the user may want to speed the imaging process by providing initial coefficient values based on what the user expects to find in the interior of the object.
  • the parameters 124 used to model the interior of the object 102 are processed by the computer 120 under the control of a set of software 130 which has three major components.
  • a "forward computation" module 132 computes the intensity of light that would emerge from the object at the exit points if the interior of the object were characterized by the currently assigned values for the scattering and attenuation coefficients.
  • ERROR_FCN is the error function and grad(p i ) is the partial derivative of the error function with respect to parameter pi.
  • the parameter updating module generates new values for the parameters by adjusting the current values of the parameters in a direction opposite to the gradient of the error function. That is, each individual parameter is adjusted in a way that is related to the multi-dimensional slope and the differences between the computed and measured intensity values:
  • P(x) i+1 P(x) i - f(ERROR_FCN,gradient, x)
  • p (x) i is the current value of one individual parameter
  • P(x)i+1 is the adjusted value of that parameter
  • f (ERROR_FCN,gradient,j) is a computed adjustment factor that is a function of the error function ERROR_FCN, the gradient of the error function ERROR_FCN, and the parameter p(x) that is being modified.
  • a successive approximation control software module 136 makes repeated calls to the forward computation and parameter update modules 132 and 134 until the value of the error function falls below a specified threshold value.
  • the final values of the scattering and attenuation coefficients from this process are then mapped by image generation software 140 so as to generate an image of the interior of the object which is displayed or printed by an image display device 142.
  • Each volume element 162 encloses a rectangular or cube-shaped portion of the area inside the measurement subsystem 110.
  • each voxel is modeled as having a set of light transmitting characteristics.
  • the light transmitting characteristics of each voxel can be represented by an attenuation coefficient, and a set of three scattering coefficients: a forward scattering coefficient, a backward scattering coefficient, and a sideways scattering coefficient.
  • Attenuation is the process by which the radiation being used for imaging is absorbed by the object. For instance, photons are converted into other forms of energy and thus are absorbed by the object.
  • Scattering is the process by which the radiated particles (e.g., photons) that are not absorbed are deflected from a straight path.
  • the attenuation coefficient for a voxel represents the probability that a radiated particle (photon) which arrives at the voxel will not emerge from it.
  • the backwards scattering coefficient for a voxel represents the probability that a radiated particle (photon) which emerges from the voxel will be reflected back to the voxel from which it arrived.
  • the sideways scattering coefficient for a voxel represents the probability that a radiated particle (photon) emerging from the voxel will have been deflected from a straight path though the voxel to a neighboring voxel.
  • the forward scattering coefficient of a voxel represents the probability that a radiated particle (photon) that emerges from the voxel will keep moving in the direction that it entered the voxel.
  • the present invention can utilize much lower radiation energies than that used in x-ray systems. As a result, the present invention can be used in situations where the amount of energy deposited in a patient or other object is of primary concern, as well as in other routine clinical procedures. Energy deposited on the object at one point of the periphery by a collimated beam of photons will not, in general, exit on the opposite side of the object because a certain amount of attenuation and scattering will occur as photons travel through the object. Instead, the path followed by the photons in the collimated beam will depend on the internal properties of the object that is being imaged and will not be known a priori.
  • a photon will travel from voxel to voxel.
  • the photon may travel forward to voxel 164, be reflected to voxel 166, or scattered sideways to a neighboring voxel 168 or 170, with this process continuing until the photon is either absorbed or emerges from the object and is detected by a sensor in the measurement subsystem.
  • FIG. 4 there is shown a somewhat more detailed block diagram of the measurement subsystem 110.
  • a two dimensional slice through the object (not shown) is modelled as having an 8x8 array 160 of voxels.
  • the size of each voxel corresponds to the distance between adjacent sensors points in the system's array of light sensors.
  • collimated beam generator 180 also called a light source
  • a light sensor 182 located at both ends of each row and column of the array 160.
  • Each generator-sensor pair is coupled by an addressable interface 184 and a communications bus 186 to the measurement subsystem's controller 190.
  • Controller 190 sequentially activates the collimated beam generators 180 one at a time, reads the measurement values from the sensors 182, and stores the measurement values in data storage module 116. In this example, since there are thirty-two light sources 180 and sensors 182, there will be thirty-two sets of measurements, each containing thirty-two measurement values.
  • the methodology of the present invention will be described primarily for a two dimensional imaging system. This is done only for ease of exposition.
  • the methodology of the present invention is equally applicable to a three dimensional imaging system, although it requires more computations because of the increased number of measurements and the increased number of parameters to be computed. In fact, only the three dimensional case is physically correct, since the scattering process inevitably scatters many particles outside the imaging plane in a two dimensional imaging system, thereby rendering the two dimensional picture inadequate.
  • the object is divided into an N by N array of volume elements, called voxels.
  • a parameter associated with a particular voxel will be denoted as X ij , with indices i and j having values between 1 and N.
  • Each voxel in the model of the object is modelled as having an attenuation coefficient V ij , and scattering coefficients f ij , b ij and S ij .
  • w ij which is equal to 1-V ij , is the probability that a photon will not be absorbed by the voxel.
  • w ij f ij is the probability that a photon entering the voxel will be transmitted through the voxel in the same direction that it entered the voxel.
  • w ij b ij is the probability that a photon will be reflected by the voxel back to the voxel from whence it came.
  • w ij s ij is the probability that a photon entering the voxel will be scattered to the side.
  • coefficients comprise a model for the "photon dynamics" of a voxel, i.e., the microscopic process by which a photon enters and leaves a given voxel. These coefficients are given in terms which are, for the most part, not directly available for measurement.
  • the present invention provides a mechanism for relating these internal voxel coefficients to a set of "exit probabilities" which correspond to light intensities that can be measured by the light sensors outside the object.
  • the set of 4N 2 variables for values of i and j in the range 1 i,j N, includes 4N external variables: for 1 i N (Eg.4) with the remaining 4N 2 - 4N variables being called
  • X is a vector consisting of the 4N detector variables as defined above.
  • Y is a vector consisting of the 4N external variables, and 2 is a vector consisting of the 4N 2 - 4N internal variables.
  • Matrix D is a square matrix which relates each of the external variables to a corresponding one of the detector variables. Therefore all of the diagonal elements of the D matrix are equal to w ij b ij , where w ij and b ij are the attenuation and backward scattering coefficients for the corresponding voxels. In addition, for each voxel located at a convex corner of the object being imaged, there is one off-diagonal element in the D matrix with a nonzero value. All the other elements of the D matrix are equal to zero.
  • the C matrix specifies the relationship between each external variable and the internal variables. At most, three elements in each row of matrix C will have a nonzero value. To understand this, remember that each variable can be expressed as a weighted combination of four other variables, corresponding to the four neighboring voxels. Looking at Figure 4, it can be seen that in the case of an external variable for a particular voxel on the periphery of the object being imaged, the equation for the variable will involve one detector variable, and three internal variables if the voxel is not on a "corner" of the object. For the few voxels located on a corner, the equation for the external variable contains only two internal variables, and two detector variables (at least one of which will have a value of zero). Therefore the rows of the C matrix for non-corner external variables have three nonzero elements and the rows of the C matrix for corner external variables have two nonzero elements.
  • Matrix A is similar to matrix C, except that it is used to relate each internal variable to other ones of the internal variables. Therefore all the diagonal elements of matrix A will be equal to 1 and each row will have at most four nonzero off-diagonal elements, namely
  • Matrix B has rows with nonzero elements only for rows corresponding to voxels along the periphery of the object being imaged. For those voxels, two elements per row may have a nonzero value for voxels on convex corners of the object, and only one element per row may have a nonzero value for other voxels on the periphery of the object.
  • the forward solution refers to the following scenario: if we know the values of w ij , f ij , b ij and s ij for all the voxels in the object, and we shine a known amount of light at a specified voxel on the periphery of the object, we can compute the amount of light that will be detected by each of the detectors around the object.
  • the forward solution is the process of computing the values of the 4N external variables represented by vector Y. Equation 9 above represents the forward solution.
  • Equation x defines the input-output relationship of light traveling through the object in terms of a 4N by 4N matrix: CA -1 B - D.
  • the entries of the CA -1 B - D matrix depend on the quantities w ij f ij , w ij b ij and W ij S ij . Equivalently, one can consider the quantities of interest to be w ij , f ij and b ij , because s ij is determined by the relation
  • CA -1 B - D Example of Forward Solution.
  • the following is a detailed example of the exact equations for an object represented by a 2 by 2 set of voxels. This example is intended to show how all the elements of the matrices A, B, C and D are defined in terms of the w ij , f ij , s ij and b ij attenuation and scattering coefficients. The reason for using such a small example is so that all the elements of each matrix can be shown on a single page.
  • a 2 by 2 array of voxels has 16 (4N 2 ) equations defining the relationships between the external , internal and detector variables :
  • the other eight equations involve internal variables.
  • the eight detector variables are the variables with indices of i or j equal to either 0 or 3.
  • the elements in the Y vector are the external variables corresponding to the voxels selected by going around the object's perimeter in a clockwise fashion, starting at position (1,1).
  • the elements of the X vector are the detector variables for detectors selected by going around the detector array in a clockwise fashion, starting at position (0,1).
  • the elements of the Z vector can be ordered in any arbitrary manner, but preferably with the variables for each voxel located close to one another.
  • the "inverse solution” is a method of solving the following problem: given a set of light intensity measurements obtained by using an array of detectors as shown in Figure 4, compute the w ij , f ij , b ij and s ij coefficients for all the voxels in the object being imaged.
  • the inverse solution is the solution to the primary problem addressed by the present invention.
  • the inverse solution is an iterative process that works as follows. Referring to Figure 5, one first takes the required measurements (box 402) and then makes an initial guess as to the values of the w ij , f ij , b ij and S ij coefficients for all the voxels in the object (box 404). If the initial guess is based on the expected structure inside the object, that may substantially improve the speed with which the inverse solution is computed.
  • the imaging system computes the expected light intensities at each of the detectors positions (box 408). In other words, one uses the above described "forward solution" to compute how much light would emerge from the object at each detector position if the object's interior were accurately modelled by the current values of the attenuation and scattering coefficients.
  • the parameter update module 134 (see Figure 1) the differences are subjected to a requirement that the sum of the squares of the differences between the measured and computed values be less than a specified threshold value. More specifically, it first computes an error function, ERROR FCN as follows:
  • the parameter update module then computes a multi-dimensional slope or gradient of this error function (box 406).
  • ERROR_FCN is the error function
  • grad(p i ) is the partial derivative of the error function with respect to parameter p i .
  • the parameters here are the 3N coefficients W ij , f ij and b ij described above. As is known to those skilled in the art of scientific computing, such partial derivatives can be numerically computed quite easily.
  • the parameter updating module generates new values for the parameters (i.e., the attenuation and scattering coefficients) (box 406) by adjusting the current values of the parameters in a direction opposite to the gradient of the error function. That is, each individual parameter is adjusted in a way that is related to the multi-dimensional slope and the differences between the computed and measured intensity values:
  • step 406 produces an updated set of coefficients that will better match the actual light propagation characteristics of the object being imaged.
  • the forward solution is recomputed as described above (box 408).
  • the error function ERROR_FCN is computed once again, and its value is compared with a threshold value to determine if another iteration of the parameter updating process is required (box 410).
  • steps 406 through 410 continues until the value of the error function falls below the specified threshold - i.e., until the computed coefficients produce a forward solution that corresponds closely to the measured intensity values.
  • a series of images is generated from the computed coefficient values. For instance, the system will typically draw an image in which the darkness or color of each point corresponds to the attenuation coefficient w ij of the corresponding voxel, and another image in which the darkness or color of each point corresponds to the sideways scattering coefficient S ij . Images based on the backwards and forward scattering coefficients b ij and f ij are also generated to depict different characteristics of the object.
  • Figure 6 a map of the attenuation characteristics of an object (i.e., in a two dimensional slice through the object).
  • Figure 7 is a map of the computed V ij coefficients (equal to 1-w ij ) for the object.
  • the one portion of the object that is not as accurately reconstructed is a "dark region" located between two sections of high attenuation materials and towards the center of the object. This is to be expected, since once the light entering the object is highly attenuated, some or most of the information regarding deeper regions of the object (i.e., behind the high attenuation regions) will be lost. It is believed by the inventors that a more accurate reconstruction would be obtained when using a three dimensional implementation of the invention. It is also expected that this problem regarding high attenuation regions may be overcome by using measurements based on two or more light frequencies that have distinct attenuation characteristics.
  • the image generation software 140 of the present invention generates additional images by mapping the computed f ij , b ij and S ij coefficients.
  • the method of mapping these coefficients into images is the same as for mapping the array of data values obtained when using CAT scan systems, magnetic resonance imaging systems, and so on.
  • the images based on the computed f ij , b ij and s ij coefficients will provide information not available from CAT scan systems, which measure only the attenuation of x-rays by an object (corresponding to the w ij coefficient which is also computed by the present invention).
  • Figure 8A shows the attenuation characteristics of one slice (i.e., a two dimensional section) of this object
  • Figures 9A and 10A show the forwards and side scattering characteristics of the object.
  • M'i d ,i s " Mi d ,i s * (1 + 0.01* ⁇ *X) (Eq.15)
  • M id , i s is the value of the measurement that would be obtained if there were no noise
  • X is a random variable taking values between 0 and 1
  • M'i d ,i s the measurement value including noise.
  • Figures 8B, 9B, and 10B represent the attenuation, forward scattering and side scattering characteristics of the reconstructed image when the measurement values include a 1% noise factor.
  • Figures 8C, 9C, and 10C represent the attenuation, forward scattering and side scattering characteristics of the reconstructed image when the measurement values include a 5% noise factor.

Abstract

An imaging system (100) generates images of the interior of an object using radiation which is attenuated and scattered by the interior of the object. The object is radiated at a number of points, and the radiation emerging from the object is measured at an array of exit points (402). The object is modelled as an array of volume elements (voxels), each voxel being assigned initial scattering and attenuation coefficients (404). The intensity of the radiation which would emerge at the exit points in an object having the assigned coefficients is calculated (408) and compared to the measured values to compute an error function (410). The error function is then minimized by modifying the values of the modelled coefficients (406) in an iterative manner until the error function falls below a specified threshold value. The final scattering and attenuation coefficients are mapped so as to generate a series of images of the interior of the object (412).

Description

IMAGING USING SCATTERED AND DIFFUSED RADIATION
The present invention relates generally to systems which produce images of the interior of an object and particularly to methods and systems for imaging the interior of an object using scattered radiation or scattered particles.
BACKGROUND OF THE INVENTION
Most systems which produce images of the interior of an object, such as the human body, utilize in-line unscattered radiation, such as x-ray sources and detectors. These systems measure, for the most part, unscattered radiation and use those measurements as the input data in a mathematical procedure of reconstructing images of the x-rayed object. The measurements in x-ray imaging systems correspond only to the relative loss of straight-through radiation as it passes through the body that is being imaged. It may be noted that magnetic resonance imaging uses an entirely different data collection and image reconstruction technique which is not related to the present invention.
Unlike the prior art, the present invention uses radiation, such as photons, phonons, neutrons or other particles, which will be scattered to a significant degree by the internal structures in various objects. The present invention radiates an object and measures the attenuated and scattered radiation at a number of points along the exterior of the object. The inventors have found that these measurements are sufficient to determine the scattering and attenuation properties of the various regions inside the object.
In the following description, since the invention can be implemented using a number of different types of radiation sources, we utilize the term "light" to represent whichever particle or type of radiation is being employed, whether it be infrared light, visible light, phonons, neutrons or other particles.
In accordance with the present invention, the interior of the object is modelled as an array of volume elements, herein called voxels. Each voxel in the model of the object has scattering and attenuation properties which are represented by numerical parameters that can be mapped so as to generate several images of the interior of the object.
It is believed by the inventors that the technique of the present invention has not been tried in the past because it has been assumed, at least implicitly, that the scattered light exiting an object would simply not contain sufficient information to generate an image of the interior of that object. For instance, any one photon travelling through the object may be scattered multiple times. In addition, photons travelling on a number of different paths can emerge at the same point on the exterior of the body. Thus, it would appear that the amount of light exiting an object at an array of points along the outside of the object would not contain sufficient information to uniquely identify the scattering and attenuation properties of the object's interior. However, the inventors have made the surprising discovery that there is, in fact, more than sufficient information to determine the internal properties of the object. SUMMARY OF THE INVENTION
In summary, the present invention is a system for generating an image of the interior of an object using radiation which is scattered and attenuated by the interior of the object. The imaging system includes a subsystem for sequentially radiating an object with light from an array of entry points near the exterior of the object, and for measuring the intensity of light emerging from the object at exit points near the exterior of the object each time that the object is radiated with light at a different entry point.
The interior of the object is modelled as an array of volume elements, herein called voxels. Each voxel in the model of the object has scattering and attenuation properties which are represented by numerical coefficients that can be mapped so as to generate several images of the interior of the object. After collecting the imaging data, the scattering and attenuation coefficients for all the voxels are assigned initial values, which helps to shorten the computation process.
Next, the system computes the intensity of light that would emerge from the object at the exit points if the interior of the object were characterized by the currently assigned values for the scattering and attenuation coefficients. Then, the differences between the measured light intensities and the computed light intensities are used to compute an error function related to the magnitude of the errors in the reconstruction. This "error" or "cost" function is then minimized using a gradient descent methodology, i.e., the value of the coefficients are modified so as to reduce the value of the error function.
The process of computing exiting light intensities based on the currently assigned values for the scattering and attenuation coefficients and then comparing the differences between the computed values and the measured values to generate a new approximation of the scattering and attenuation properties of the object continues until the error function (which corresponds to the differences between the measured and computed exiting light intensities) falls below a specified threshold value. The final values of the scattering and attenuation coefficients from this process are then mapped so as to generate a series of images of the interior of the object depicting the attenuation and scattering characteristics of the object's interior.
BRIEF DESCRIPTION OF THE DRAWINGS
Additional objects and features of the invention will be more readily apparent from the following detailed description and appended claims when taken in conjunction with the drawings, in which:
Figure 1 is a block diagram of an imaging system in accordance with the present invention.
Figure 2 schematically depicts how an object is modelled as an array of volume elements.
Figure 3 schematically depicts the attenuation, scattering and reflection properties of a single volume element. Figure 4 is a block diagram of a measurement subsystem in accordance with the present invention. Figure 5 is a flow chart of the imaging method used in the preferred embodiment of the invention.
Figure 6 depicts a map of the attenuation characteristics of an object.
Figure 7 depicts a map of a reconstruction of the object shown in Figure 6 generated using a prototype of the present invention.
Figures 8A-8C depict a map of the forward scattering characteristics of a three dimensional object and two maps of reconstructions of the object with varying amounts of noise in the measurement signal.
Figures 9A-9C depict a map of the attenuation characteristics of a three dimensional object and two maps of reconstructions of the object with varying amounts of noise in the measurement signal.
Figures 10A-10C depict a map of the side scattering characteristics of a three dimensional object and two maps of reconstructions of the object with varying amounts of noise in the measurement signal.
DESCRIPTION OF THE PREFERRED EMBODIMENT
Referring to Figure 1, there is shown an imaging system 100 which generates an image of the internal structure of an object 102 based on measurements of scattered radiation which passes through the object.
A measurement subsystem 110 radiates the object 102 with radiation, such as visible or infrared light, which will be scattered to a significant degree by the internal structures in the object. The measurement subsystem 110 includes a radiator system 112 (e.g., a particle radiator or a lighting system) for sequentially radiating an object with light from an array of entry points near the exterior of the object, and a sensor system 114 (e.g., an array of particle detectors or light sensors) for measuring the intensity of light emerging from the object at an array exit points near the exterior of the object each time that the object is radiated with light at a different entry point. The measurement data generated by the measurement subsystem 110 is stored in a data storage device 116 and processed by a digital computer 120. Furthermore, the computer 120 generates positioning control signals A and B for controlling the set of positions of the radiation source(s) 112 and detectors 114, respectively, so as to obtain data sets sufficient for reconstructing an image of the interior structure of the object 102.
A software model 122 is used to represent the interior of the object as an array of volume elements, called voxels. In accordance with this model 122, the interior of the object 122 is modelled by a set of numerical parameters 124 which include scattering and attenuation coefficients for each of the voxels. These coefficients are assigned initial values by either the model 122, or by a user via the user interface 126. The model 122 in the preferred embodiment initializes the coefficients in all the voxels with the same set of values, which corresponds to an object with a totally uniform interior. However, the user may want to speed the imaging process by providing initial coefficient values based on what the user expects to find in the interior of the object.
The parameters 124 used to model the interior of the object 102 are processed by the computer 120 under the control of a set of software 130 which has three major components. A "forward computation" module 132 computes the intensity of light that would emerge from the object at the exit points if the interior of the object were characterized by the currently assigned values for the scattering and attenuation coefficients.
A parameter updating module 134 computes updated parameter values based on the differences between the measured light intensities and the computed light intensities. More particularly, it first computes an error function which is equal to the sum of the squares of the differences between the measured light intensities and the computed light intensities. ERROR_FCN = (Measured Intensity at Sensor i
Figure imgf000009_0002
- Computed Intensity at Sensor i)2
It then computes a multi-dimensional slope or gradient of this error function. This multi-dimensional gradient is vector, the components of which are the partial derivatives of the error function with respect to each of the parameter values: gradient = (grad(p1), grad(p2), .... grad(p4N)) grad(Pi) = (ERROR_FCN)
Figure imgf000009_0001
where ERROR_FCN is the error function and grad(pi) is the partial derivative of the error function with respect to parameter pi.
Next, the parameter updating module generates new values for the parameters by adjusting the current values of the parameters in a direction opposite to the gradient of the error function. That is, each individual parameter is adjusted in a way that is related to the multi-dimensional slope and the differences between the computed and measured intensity values:
P(x)i+1 = P(x)i - f(ERROR_FCN,gradient, x) where p (x)i is the current value of one individual parameter, P(x)i+1 is the adjusted value of that parameter, and f (ERROR_FCN,gradient,j) is a computed adjustment factor that is a function of the error function ERROR_FCN, the gradient of the error function ERROR_FCN, and the parameter p(x) that is being modified.
A successive approximation control software module 136 makes repeated calls to the forward computation and parameter update modules 132 and 134 until the value of the error function falls below a specified threshold value. The final values of the scattering and attenuation coefficients from this process are then mapped by image generation software 140 so as to generate an image of the interior of the object which is displayed or printed by an image display device 142.
Referring to Figure 2, there is shown a two dimensional example of how the interior of an object is modelled as an array 160 of volume elements. Each volume element 162 encloses a rectangular or cube-shaped portion of the area inside the measurement subsystem 110.
Referring to Figure 3, each voxel is modeled as having a set of light transmitting characteristics. In particular, it is assumed that the light transmitting characteristics of each voxel can be represented by an attenuation coefficient, and a set of three scattering coefficients: a forward scattering coefficient, a backward scattering coefficient, and a sideways scattering coefficient. As photons travel through each voxel inside the object, they are subject to attenuation and scattering. Attenuation is the process by which the radiation being used for imaging is absorbed by the object. For instance, photons are converted into other forms of energy and thus are absorbed by the object. Scattering is the process by which the radiated particles (e.g., photons) that are not absorbed are deflected from a straight path.
The attenuation coefficient for a voxel represents the probability that a radiated particle (photon) which arrives at the voxel will not emerge from it. The backwards scattering coefficient for a voxel represents the probability that a radiated particle (photon) which emerges from the voxel will be reflected back to the voxel from which it arrived. The sideways scattering coefficient for a voxel represents the probability that a radiated particle (photon) emerging from the voxel will have been deflected from a straight path though the voxel to a neighboring voxel. Finally, the forward scattering coefficient of a voxel represents the probability that a radiated particle (photon) that emerges from the voxel will keep moving in the direction that it entered the voxel.
If one were to use radiation such as X-rays, gamma rays, or other particles which provide good transmission through the object, one could minimize the effect of scattering and thus reconstruct two or three dimensional maps of the attenuation characteristics of the object. That is the methodology used in standard x-ray tomographic reconstruction, which is a well defined and well understood technology. In the limiting case, when back scattering and side scattering are so small that they can be ignored, the present invention becomes equivalent to standard x-ray tomographic reconstruction.
The present invention can utilize much lower radiation energies than that used in x-ray systems. As a result, the present invention can be used in situations where the amount of energy deposited in a patient or other object is of primary concern, as well as in other routine clinical procedures. Energy deposited on the object at one point of the periphery by a collimated beam of photons will not, in general, exit on the opposite side of the object because a certain amount of attenuation and scattering will occur as photons travel through the object. Instead, the path followed by the photons in the collimated beam will depend on the internal properties of the object that is being imaged and will not be known a priori. If one conceives of the object as a collection of small volume units or voxels as shown in Figure 2, a photon will travel from voxel to voxel. At each voxel 162 the photon may travel forward to voxel 164, be reflected to voxel 166, or scattered sideways to a neighboring voxel 168 or 170, with this process continuing until the photon is either absorbed or emerges from the object and is detected by a sensor in the measurement subsystem.
Referring to Figure 4, there is shown a somewhat more detailed block diagram of the measurement subsystem 110. In this example, a two dimensional slice through the object (not shown) is modelled as having an 8x8 array 160 of voxels. As shown, the size of each voxel corresponds to the distance between adjacent sensors points in the system's array of light sensors.
There is a collimated beam generator 180 (also called a light source) and a light sensor 182 located at both ends of each row and column of the array 160. Each generator-sensor pair is coupled by an addressable interface 184 and a communications bus 186 to the measurement subsystem's controller 190. Controller 190 sequentially activates the collimated beam generators 180 one at a time, reads the measurement values from the sensors 182, and stores the measurement values in data storage module 116. In this example, since there are thirty-two light sources 180 and sensors 182, there will be thirty-two sets of measurements, each containing thirty-two measurement values.
COMPUTATION METHOD.
It should be noted that the methodology of the present invention will be described primarily for a two dimensional imaging system. This is done only for ease of exposition. The methodology of the present invention is equally applicable to a three dimensional imaging system, although it requires more computations because of the increased number of measurements and the increased number of parameters to be computed. In fact, only the three dimensional case is physically correct, since the scattering process inevitably scatters many particles outside the imaging plane in a two dimensional imaging system, thereby rendering the two dimensional picture inadequate.
As shown in Figure 4, the object is divided into an N by N array of volume elements, called voxels. A parameter associated with a particular voxel will be denoted as Xij, with indices i and j having values between 1 and N. Each voxel in the model of the object is modelled as having an attenuation coefficient Vij, and scattering coefficients fij, bij and Sij. wij, which is equal to 1-Vij, is the probability that a photon will not be absorbed by the voxel. wijfij is the probability that a photon entering the voxel will be transmitted through the voxel in the same direction that it entered the voxel. wijbij is the probability that a photon will be reflected by the voxel back to the voxel from whence it came. wijsij is the probability that a photon entering the voxel will be scattered to the side. The relative values of the scattering coefficients are governed by the equation fij + bij + 2*sij = 1 (Eq.1) for a two dimensional imaging system. For a three dimensional imaging system the scattering coefficients are governed by the equation fij + bij + 4*sij = 1 (Eq.2)
These coefficients comprise a model for the "photon dynamics" of a voxel, i.e., the microscopic process by which a photon enters and leaves a given voxel. These coefficients are given in terms which are, for the most part, not directly available for measurement.
The present invention provides a mechanism for relating these internal voxel coefficients to a set of "exit probabilities" which correspond to light intensities that can be measured by the light sensors outside the object.
The basis for our model is a "two step Markov transition mechanism". That is, we assume that the transition mechanism of a photon that is at voxel position (i,j) depends only on its present and immediately previous position, and not on its previous history. Given the previous position of a photon, we can say that a photon is at position (i,j), having arrived from the up(u), down(d), right(r) or left(l) directions.
For a given detector position (x,y) we will derive equations for the quantities:
Figure imgf000014_0001
and
Figure imgf000014_0002
1 i, j N where the superscripts u, d, r and 1 denote photons entering voxel (i,j) from the up, down, right and left positions, respectively.
These quantities
Figure imgf000014_0003
stand for the probabilities that a photon that has started from pixel (i,j), having arrived from one of its four neighbors, would exit the object at a specified one of the detectors outside the object. For an object modelled by an NxN array of voxels, there are 4N2 such quantities for each specified detector. It is important to remember that each quantity
Figure imgf000015_0004
corresponds to a particular detector position, even though our notation for these probability quantities
Figure imgf000015_0003
f does not denote the detector position. If detector positions were given by (x,y) the probability quantities could be denoted as (x,y).
Figure imgf000015_0002
It should also be noted that most of these probability values
Figure imgf000015_0005
cannot be directly measured, since we only inject photons at the boundary of the object and thus can measure only the ones that correspond to voxels on the border of the object, namely:
for the top border
for the right border
for the lower border
Figure imgf000015_0001
for tne left border
Of the total 4N2 such quantities that we have introduced for each light sensor, we can directly measure only the 4N quantities noted above. We will refer to these 4N quantities as "external" or "source" variables.
In addition, for reasons that will be clear from the following analysis, we augment the NxN array of voxels with two rows and two columns of N pixels at locations 0,1 through 0,N for the top border
1,N+1 through N,N+1 for the right border
N+1,1 through N+1,N for the bottom border
1,0 through N,0 for the left border as shown in Figure 4. In conjunction with these 4N extra voxels, we have 4N quantities denoted as
Figure imgf000016_0006
for the values of the indices (i,j) listed above. These variables, herein called "detector" variables, will serve as "markers" that indicate the detector for which the probability values are being computed.
For the top border we will give values to
Figure imgf000016_0003
P ,j, for the right border to
Figure imgf000016_0005
for the left border to p
Figure imgf000016_0004
and for the bottom border we will give values to
Figure imgf000016_0008
If the detector in question is, say, at location (0,3), we assign to
Figure imgf000016_0007
the value 1, and we assign the value 0 to all the other 4N-1 detector variables. Thus these variables serve as markers to indicate which detector is active.
The remaining
Figure imgf000016_0002
variables, for i or j equal to 0 or N+1 do not enter into the equations shown below and therefore need not be assigned any values.
The basic matrix equation relating the probability variables to the attenuation and scattering coefficients is as follows:
Figure imgf000016_0001
Note that there are one of the above matrix equations for each of the N2 voxels. Each of these matrix equations is really four scalar equations. Thus, for a single selected detector we have 4N2 equations. Since there are 4N detectors, we will have a total of 4N * 4N2 = 16N3 equations.
As an example, the following are the equations, as defined by the above matrix equation, for voxels (i=1, j=3) and (i=1, j=2) for a detector at position (0,3). p >
f
Figure imgf000017_0001
Note that in the above equations for a detector at position
(0,3),
Figure imgf000017_0004
equals 1 and
Figure imgf000017_0005
equals 0. The set of 4N2 variables
Figure imgf000017_0003
, for values of i and j in the range 1 i,j N, includes 4N external variables:
Figure imgf000017_0002
for 1 i N (Eg.4) with the remaining 4N2 - 4N variables being called
Figure imgf000017_0006
"internal" variables.
We now define three vectors X, Y, and Z as follows. X is a vector consisting of the 4N detector variables as defined above. Y is a vector consisting of the 4N external variables, and 2 is a vector consisting of the 4N2 - 4N internal variables. Next, recalling that there are 4N2 equations associated with each detector, we write those 4N2 equations in matrix form as follows:
#Rows:
Figure imgf000018_0001
(Eq.5)
#Columns: 4N 4N2 - 4N 4N where I is the identity matrix, C is a matrix having 4N rows and 4N2 - 4N columns, D is a 4N by 4N matrix, A is a square matrix having 4N2 - 4N rows and 4N2 - 4N columns, and B is a matrix having 4N2 - 4N rows and 4N columns.
In addition, recall that selecting one detector amounts to setting one of the components of the detector variables in the X vector equal to 1, and setting all the other detector variables in the X vector to 0. Thus there will be 4N such sets of equations, one for each of the 4N detectors.
An equivalent representation of the above matrix equation is as follows:
Y + CZ + DX = 0 (Eq.6) AZ + BX = 0 (Eq.7)
Matrix D is a square matrix which relates each of the external variables to a corresponding one of the detector variables. Therefore all of the diagonal elements of the D matrix are equal to wijbij, where wij and bij are the attenuation and backward scattering coefficients for the corresponding voxels. In addition, for each voxel located at a convex corner of the object being imaged, there is one off-diagonal element in the D matrix with a nonzero value. All the other elements of the D matrix are equal to zero.
The C matrix specifies the relationship between each external variable and the internal variables. At most, three elements in each row of matrix C will have a nonzero value. To understand this, remember that each variable can be expressed as a weighted combination of four other variables, corresponding to the four neighboring voxels. Looking at Figure 4, it can be seen that in the case of an external variable for a particular voxel on the periphery of the object being imaged, the equation for the variable will involve one detector variable, and three internal variables if the voxel is not on a "corner" of the object. For the few voxels located on a corner, the equation for the external variable contains only two internal variables, and two detector variables (at least one of which will have a value of zero). Therefore the rows of the C matrix for non-corner external variables have three nonzero elements and the rows of the C matrix for corner external variables have two nonzero elements.
Matrix A is similar to matrix C, except that it is used to relate each internal variable to other ones of the internal variables. Therefore all the diagonal elements of matrix A will be equal to 1 and each row will have at most four nonzero off-diagonal elements, namely
-wijfij, -wijbij, -wijsij, -wijsij .
Since the sum of these off-diagonal elements is -wij, and 0 ≤ wij ≤ 1, we have a diagonally dominant matrix. If we can impose the physically reasonable restriction wij < 1, matrix A is strictly diagonally dominant. In this case, matrix A is invertible. Matrix B has rows with nonzero elements only for rows corresponding to voxels along the periphery of the object being imaged. For those voxels, two elements per row may have a nonzero value for voxels on convex corners of the object, and only one element per row may have a nonzero value for other voxels on the periphery of the object.
By inverting the A matrix, the above matrix equations (Eq.
7 and Eq.8) can be rewritten as follows:
Z = -A-1BX (Eq.8)
Y = -CZ - DX - (CA-1B - D)X (Eq.9)
THE FORWARD SOLUTION.
The term "the forward solution" refers to the following scenario: if we know the values of wij, fij, bij and sij for all the voxels in the object, and we shine a known amount of light at a specified voxel on the periphery of the object, we can compute the amount of light that will be detected by each of the detectors around the object. In other words, the forward solution is the process of computing the values of the 4N external variables represented by vector Y. Equation 9 above represents the forward solution. Given any value choice of values for vector X (i.e., with one element set to 1 and all the other elements set to 0), Equation x defines the input-output relationship of light traveling through the object in terms of a 4N by 4N matrix: CA-1B - D.
The entries of the CA-1B - D matrix depend on the quantities wijfij, wijbij and WijSij. Equivalently, one can consider the quantities of interest to be wij, fij and bij, because sij is determined by the relation
Sij =
Figure imgf000020_0001
(1 - fij - bij) (Eq.10) for a two dimensional object and
Sij =
Figure imgf000021_0002
(1 - fij - bij) (Eq.11) for a three dimensional object. In summary, for a two dimensional object there are 3N2 variables of the form wij, fij and bij which represent the light transmission characteristics of the object being imaged, and there are 16N2 equations relating these variables and given by entries of the 4N by 4N matrix
CA-1B - D. Example of Forward Solution. The following is a detailed example of the exact equations for an object represented by a 2 by 2 set of voxels. This example is intended to show how all the elements of the matrices A, B, C and D are defined in terms of the wij, fij, sij and bij attenuation and scattering coefficients. The reason for using such a small example is so that all the elements of each matrix can be shown on a single page.
A 2 by 2 array of voxels has 16 (4N2) equations defining the relationships between the external , internal and detector variables :
] f / b p p
Figure imgf000021_0001
p p
Figure imgf000022_0001
(Eq.12)
The arrows above point to the eight equations involving external variables. The other eight equations involve internal variables. The eight detector variables are the variables
Figure imgf000022_0002
with indices of i or j equal to either 0 or 3.
In terms of the matrix equation above, and using the short hand notation
Fij = wijfij
Bij = wijbij
Sij = wijsij X, Y, Z, A, B, C and D are defined as follows:
Figure imgf000023_0001
Figure imgf000024_0001
Note that the elements in the Y vector are the external variables corresponding to the voxels selected by going around the object's perimeter in a clockwise fashion, starting at position (1,1). The elements of the X vector are the detector variables for detectors selected by going around the detector array in a clockwise fashion, starting at position (0,1). The elements of the Z vector can be ordered in any arbitrary manner, but preferably with the variables for each voxel located close to one another.
THE INVERSE SOLUTION.
The "inverse solution" is a method of solving the following problem: given a set of light intensity measurements obtained by using an array of detectors as shown in Figure 4, compute the wij, fij, bij and sij coefficients for all the voxels in the object being imaged. In other words, the inverse solution is the solution to the primary problem addressed by the present invention.
Generally speaking, the inverse solution is an iterative process that works as follows. Referring to Figure 5, one first takes the required measurements (box 402) and then makes an initial guess as to the values of the wij, fij, bij and Sij coefficients for all the voxels in the object (box 404). If the initial guess is based on the expected structure inside the object, that may substantially improve the speed with which the inverse solution is computed. However, for the purposes of testing the first prototypes of the invention, the inventors have used an initial set of values corresponding to an object of uniform qualities, such as Wij = 0.5, fij = 0.7, bij = 0, and Sij = 0.1 for all voxels inside the object. Next, given the current values of the attenuation and scattering coefficients, the imaging system computes the expected light intensities at each of the detectors positions (box 408). In other words, one uses the above described "forward solution" to compute how much light would emerge from the object at each detector position if the object's interior were accurately modelled by the current values of the attenuation and scattering coefficients. Then, one determines the differences between the computed detector intensity values and those measured in step 402. These differences are then compared with a threshold to determine whether further computations are required (box 410). In the simplest embodiment, further computations are required if any of the differences (i.e., between a any computed intensity value and the corresponding measurement) exceed a threshold value.
In the preferred embodiment, the parameter update module 134 (see Figure 1) the differences are subjected to a requirement that the sum of the squares of the differences between the measured and computed values be less than a specified threshold value. More specifically, it first computes an error function, ERROR FCN as follows:
ERROR_FCN = (Measured Intensity at Sensor i
- Computed Intensity at Sensor i)2
Figure imgf000025_0001
(Eq.13) where S is the number of sensors at which measurements are taken. In the ideal case for a two dimensional system, S will be equal to 4N for an N by N array of voxels. The parameter update module then computes a multi-dimensional slope or gradient of this error function (box 406). This multi-dimensional gradient is a vector, the components of which are the partial derivatives of the error function with respect to each of the parameter values: gradient = (grad(p1), grad(p2), .... grad(p4N)) grad(pi) = | (ERROR_FCN)
Figure imgf000026_0001
where ERROR_FCN is the error function and grad(pi) is the partial derivative of the error function with respect to parameter pi. Note that the parameters here are the 3N coefficients Wij, fij and bij described above. As is known to those skilled in the art of scientific computing, such partial derivatives can be numerically computed quite easily.
Next, the parameter updating module generates new values for the parameters (i.e., the attenuation and scattering coefficients) (box 406) by adjusting the current values of the parameters in a direction opposite to the gradient of the error function. That is, each individual parameter is adjusted in a way that is related to the multi-dimensional slope and the differences between the computed and measured intensity values:
P(x)i+1 - P(x)i - f(ERROR_FCN,gradient,x) (Eq.14) where p(x)i is the current value of one individual parameter, p(x)i+1 is the adjusted value of that parameter, and f (ERROR_FCN,gradient,j) is a computed adjustment factor that is a function of the error function ERROR_FCN, the gradient of the error function ERROR_FCN, and the parameter p(x) that is being modified. There are a number of well known numerical techniques for minimizing multidimensional (i.e., multivariable) functions, which we herein call "gradient descent" algorithms. See, for example, William H. Press, Brian P. Flannery, Saul A. Teukolsky, William T. Vetterling, Numerical Recipes, The Art of Scientific Computing, Cambridge University Press, sections 10.5 through 10.10, pp.294-334 (1976), which is hereby incorporated by reference. Any one of a number of such methods can be used with the present invention. The particular method used in preferred embodiment is a conjugate gradient method. Using this methodology, the values of the coefficients are changed so that the error function is minimized during most iterations. However, as will be understood by those skilled in the art, to prevent the process from getting stuck in local minima, the process occasionally allows a random move that may increase the value of the error function. In summary, regardless of which gradient descent method is used, step 406 produces an updated set of coefficients that will better match the actual light propagation characteristics of the object being imaged. Once the new coefficient values have been computed (box 406), the forward solution is recomputed as described above (box 408). Then the error function ERROR_FCN is computed once again, and its value is compared with a threshold value to determine if another iteration of the parameter updating process is required (box 410).
The iterative process represented by steps 406 through 410 continues until the value of the error function falls below the specified threshold - i.e., until the computed coefficients produce a forward solution that corresponds closely to the measured intensity values. Finally, once the iterative computation process of steps 406 through 410 is completed, a series of images is generated from the computed coefficient values. For instance, the system will typically draw an image in which the darkness or color of each point corresponds to the attenuation coefficient wij of the corresponding voxel, and another image in which the darkness or color of each point corresponds to the sideways scattering coefficient Sij. Images based on the backwards and forward scattering coefficients bij and fij are also generated to depict different characteristics of the object.
Referring to Figures 6 and 7, there is shown in Figure 6 a map of the attenuation characteristics of an object (i.e., in a two dimensional slice through the object). Using a simulated prototype of the present invention, the attenuation characteristics of this object were reconstructed, resulting in the map shown in Figure 7. More specifically, Figure 7 is a map of the computed Vij coefficients (equal to 1-wij) for the object.
It is interesting to note that while the reconstructed image in Figuire 7 is generally quite accurate, the one portion of the object that is not as accurately reconstructed is a "dark region" located between two sections of high attenuation materials and towards the center of the object. This is to be expected, since once the light entering the object is highly attenuated, some or most of the information regarding deeper regions of the object (i.e., behind the high attenuation regions) will be lost. It is believed by the inventors that a more accurate reconstruction would be obtained when using a three dimensional implementation of the invention. It is also expected that this problem regarding high attenuation regions may be overcome by using measurements based on two or more light frequencies that have distinct attenuation characteristics. The image generation software 140 of the present invention generates additional images by mapping the computed fij, bij and Sij coefficients. The method of mapping these coefficients into images is the same as for mapping the array of data values obtained when using CAT scan systems, magnetic resonance imaging systems, and so on. However, it is expected that the images based on the computed fij, bij and sij coefficients will provide information not available from CAT scan systems, which measure only the attenuation of x-rays by an object (corresponding to the wij coefficient which is also computed by the present invention). Next we show a set of images generated using a simulation of the invention with a three dimensional object modelled as a three dimensional array of voxels. Figure 8A shows the attenuation characteristics of one slice (i.e., a two dimensional section) of this object, and Figures 9A and 10A show the forwards and side scattering characteristics of the object.
In these simulations, the measurements upon which the calculations are performed have been modified by adding different levels of multiplicative noise. More precisely, noise was added to the measurements as follows. Let Mid, is denote a measurement by a detector at position id when using a radiation source at position is. In other words. Mid, is is the measurement value that would be obtained if there were no noise. Then we introduce a multiplicative noise of α% as follows:
M'id,is " Mid,is * (1 + 0.01*α*X) (Eq.15) where Mid, is is the value of the measurement that would be obtained if there were no noise, X is a random variable taking values between 0 and 1, and M'id,is the measurement value including noise.
Using two sets of modified measurement values, one set of measurement values representing a 1% noise factor and a section set representing a 5% noise factor, the attenuation and scattering coefficients for the object were generated using the above described inverse solution technique. Figures 8B, 9B, and 10B represent the attenuation, forward scattering and side scattering characteristics of the reconstructed image when the measurement values include a 1% noise factor. Figures 8C, 9C, and 10C represent the attenuation, forward scattering and side scattering characteristics of the reconstructed image when the measurement values include a 5% noise factor. As can be seen, while the reconstructed images generated are affected by noise, the general characteristics of the object are still reconstructed quite well with a 5% noise factor.
While the present invention has been described with reference to a few specific embodiments, the description is illustrative of the invention and is not to be construed as limiting the invention. Various modifications may occur to those skilled in the art without departing from the true spirit and scope of the invention as defined by the appended claims.

Claims

WHAT IS CLAIMED IS:
1. A system for generating images corresponding to the internal characteristics of an object, comprising:
radiating means for generating a directed beam of radiation, said beam being directed at distinct times at each of a multiplicity of input positions around an object; sensor means arrayed around at least a portion of said body for measuring the intensity of radiation emerging from said body at a multiplicity of sensor positions;
data storage means for storing a multiplicity of sets of measured intensity values, each set of intensity values corresponding to the intensities of radiation measured by said sensor means while said radiating means is generating a beam directed at distinct ones of said input positions; modeling means for modeling the interior of said object as an array of voxels, each voxel being characterized by an attenuation coefficient and a plurality of scattering coefficients, said coefficients corresponding to portions of radiation entering said voxel which are absorbed and scattered by said voxel;
said modeling means including:
coefficient denoting means for denoting current values for said attenuation and scattering coefficients for all of said voxels in said model of the interior of said object, including means for assigning initial values to said coefficients;
forward computing means for computing a set of intensity values comprising the intensity of radiation that would emerge from said object at said sensor positions while said object is radiated at each of said input positions, were said object characterized by said current values for said attenuation and scattering coefficients as denoted by said coefficient denoting means; and
updating means, coupled to said forward computing means, said data storage means and said coefficient denoting means, for evaluating differences between said computed and measured intensity values, and for generating new values for said attenuation and scattering coefficients which reduce the differences between said computed and measured intensity values; and
control means, coupled to said forward computing and updating means, including means for performing a successive approximation process in which said forward computing and updating means repeatedly compute intensity values and then update said attenuation and scattering coefficients until the differences between said computed and measured intensity values meet predefined criteria.
2. An imaging system as set forth in Claim 1, further comprising:
image generation means for generating an image corresponding to at least a subset of said attenuation and scattering coefficients.
3. An imaging system as set forth in Claim 1, said updating means including:
gradient means for computing a predefined function of said differences between said computed and measured intensity values and for computing a multi-dimensional slope corresponding to partial differentials of said predefined function with respect to each of said coefficients;
means for generating new values for said coefficients coefficients by adjusting the current values of said coefficients by amounts corresponding to said multi-dimensional slope.
4. An imaging system as set forth in Claim 1, said updating means including:
gradient means for computing a predefined function corresponding to the sum of the squares of said differences between said computed and measured intensity values, and for computing a multi-dimensional slope corresponding to partial differentials of said predefined function with respect to each of said coefficients;
means for generating new values for said coefficients coefficients by adjusting the current values of said coefficients by amounts corresponding to said multi-dimensional slope.
5. A system for generating images corresponding to the internal characteristics of an object, comprising:
light generating means for generating a directed beam of light, said beam being directed at distinct times at each of a multiplicity of input positions around an object; sensor means arrayed around at least a portion of said body for measuring the intensity of light emerging from said body at a multiplicity of sensor positions;
data storage means for storing a multiplicity of sets of measured intensity values, each set of intensity values corresponding to the intensities of light measured by said sensor means while said light generating means is generating a beam directed at distinct one of said input positions;
modeling means for modeling the interior of said object as an array of voxels, each voxel being characterized by an attenuation coefficient and a plurality of scattering coefficients, said coefficients corresponding to portions of light entering said voxel which are absorbed and scattered by said voxel;
said modeling means including:
coefficient denoting means for denoting current values for said attenuation and scattering coefficients for all of said voxels in said model of the interior of said object, including means for assigning initial values to said coefficients;
forward computing means for computing a set of intensity values comprising the intensity of radiation that would emerge from said object at said sensor positions while said object is radiated with light at each of said input positions, were said object characterized by said current values for said attenuation and scattering coefficients as denoted by said coefficient denoting means; and
updating means, coupled to said forward computing means, said data storage means and said coefficient denoting means, for evaluating differences between said computed and measured intensity values, and for generating new values for said attenuation and scattering coefficients which reduce the differences between said computed and measured intensity values; and
control means, coupled to said forward computing and updating means, including means for performing a successive approximation process in which said forward computing and updating means repeatedly compute intensity values and then update said attenuation and scattering coefficients until the differences between said computed and measured intensity values meet predefined criteria.
6. An imaging system as set forth in Claim 5, further comprising:
image generation means for generating an image corresponding to at least a subset of said transmission, reflection and scattering coefficients.
7. An imaging system as set forth in Claim 5, said updating means including:
gradient means for computing a predefined function of said differences between said computed and measured intensity values and for computing a multi-dimensional slope corresponding to partial differentials of said predefined function with respect to each of said coefficients;
means for generating new values for said coefficients coefficients by adjusting the current values of said coefficients by amounts corresponding to said multi-dimensional slope.
8. An imaging system as set forth in Claim 5, said updating means including:
gradient means for computing a predefined function corresponding to the sum of the squares of said differences between said computed and measured intensity values, and for computing a multi-dimensional slope corresponding to partial differentials of said predefined function with respect to each of said coefficients;
means for generating new values for said coefficients coefficients by adjusting the current values of said coefficients by amounts corresponding to said multi-dimensional slope.
9. A method of imaging the interior of an object, the steps of the method comprising:
generating a directed beam of radiation directed at distinct times at each of a multiplicity of input positions around an object;
measuring the intensity of radiation emerging from said body at a multiplicity of sensor positions and storing sets of measured intensity values, each set of intensity values corresponding to said measured intensities while said object is being radiated by a beam directed at distinct ones of said input positions;
modeling the interior of said object as an array of voxels, each voxel being characterized by an attenuation coefficient and a plurality of scattering coefficients, said coefficients corresponding to portions of radiation entering said voxel which are absorbed and scattered by said voxel;
said modeling step including the steps of:
denoting current values for said attenuation and scattering coefficients for all of said voxels;
computing a set of intensity values comprising the intensity of radiation that would emerge from said object at said sensor positions while said object is radiated at each of said input positions, were said object characterized by said current values for said attenuation and scattering coefficients;
evaluating differences between said computed and measured intensity values, and
updating said current values of said coefficients by computing new values for said attenuation and scattering coefficients which reduce the differences between said computed and measured intensity values; and performing a successive approximation process by repeatedly performing said computing, evaluating and updating steps until the differences between said computed and measured intensity values meet predefined criteria.
10. A method of imaging the interior of an object as set forth in Claim 9, further comprising the step of:
generating an image corresponding to at least a subset of said attenuation and scattering coefficients.
11. A method of imaging the interior of an object as set forth in Claim 10, said updating step including the steps of:
computing a predefined function of the differences between said computed and measured intensity values,
computing a multi-dimensional slope corresponding to partial differentials of said predefined function with respect to each of said coefficients; and
computing new values for said coefficients coefficients by adjusting the current values of said coefficients by amounts corresponding to said multi-dimensional slope.
12. A method of imaging the interior of an object as set forth in Claim 10, said updating step including the steps of:
computing a predefined function corresponding to the sum of the squares of the differences between said computed and measured intensity values;
computing a multi-dimensional slope corresponding to partial differentials of said predefined function with respect to each of said coefficients; and
computing new values for said coefficients coefficients by adjusting the current values of said coefficients by amounts corresponding to said multi-dimensional slope.
PCT/US1990/006664 1989-11-22 1990-11-20 Imaging using scattered and diffused radiation WO1991007655A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP91501478A JPH05502393A (en) 1989-11-22 1990-11-20 Image formation using scattered/diffuse lines

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US440,088 1989-11-22
US07/440,088 US5070455A (en) 1989-11-22 1989-11-22 Imaging system and method using scattered and diffused radiation

Publications (1)

Publication Number Publication Date
WO1991007655A1 true WO1991007655A1 (en) 1991-05-30

Family

ID=23747387

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1990/006664 WO1991007655A1 (en) 1989-11-22 1990-11-20 Imaging using scattered and diffused radiation

Country Status (3)

Country Link
US (1) US5070455A (en)
JP (1) JPH05502393A (en)
WO (1) WO1991007655A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0627698A1 (en) * 1993-06-03 1994-12-07 Hamamatsu Photonics K.K. Optical computer tomograph
EP0633527A1 (en) * 1993-07-09 1995-01-11 Hamamatsu Photonics K.K. Optical computer tomograph
EP0692708A3 (en) * 1994-07-14 1996-10-30 Hitachi Ltd Imaging method for spatial distributions of absorber concentrations
US5719398A (en) * 1995-09-11 1998-02-17 U.S. Philips Corporation Method and device for imaging the interior of a turbid medium

Families Citing this family (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU700802B2 (en) * 1993-10-29 1999-01-14 Trustees Of The University Of Pennsylvania, The Object imaging using diffuse light
JPH09500566A (en) * 1993-11-30 1997-01-21 ベル コミュニケーションズ リサーチ,インコーポレイテッド Imaging system and method using direct reconstruction of scattered radiation
US5528365A (en) * 1994-03-01 1996-06-18 The Trustees Of The University Of Pennsylvania Methods and apparatus for imaging with diffuse light
US5758653A (en) * 1995-04-10 1998-06-02 Bell Communications Research, Inc. Simultaneous absorption and diffusion imaging system and method using direct reconstruction of scattered radiation
US5787888A (en) * 1995-05-15 1998-08-04 Schotland; John Carl Absorption tomography system and method using direct reconstruction of scattered radiation
US5747810A (en) * 1995-05-15 1998-05-05 Univ. Of Pennsylvania Simultaneous absorption and diffusion tomography system and method using direct reconstruction of scattered radiation
US7328059B2 (en) * 1996-08-23 2008-02-05 The Texas A & M University System Imaging of light scattering tissues with fluorescent contrast agents
JP3819032B2 (en) * 1995-08-24 2006-09-06 ザ・テキサス・エイ・アンド・エム・ユニバーシティ・システム Imaging and spectroscopic analysis based on fluorescence lifetime in tissues and other random media
US5905261A (en) * 1995-12-01 1999-05-18 Schotland; John Carl Imaging system and method using direct reconstruction of scattered radiation
US5835617A (en) * 1996-01-18 1998-11-10 Hamamatsu Photonics K.K. Optical computer tomographic apparatus and image reconstruction method using optical computer tomography
JP3662376B2 (en) * 1996-05-10 2005-06-22 浜松ホトニクス株式会社 Internal characteristic distribution measuring method and apparatus
US5963658A (en) * 1997-01-27 1999-10-05 University Of North Carolina Method and apparatus for detecting an abnormality within a host medium
US5762607A (en) * 1997-03-19 1998-06-09 Schotland; John Carl Emission tomography system and method using direct reconstruction of scattered radiation
GB9724835D0 (en) 1997-11-25 1998-01-21 Univ Manchester Method and apparatus for detecting an object
US6064917A (en) * 1997-12-22 2000-05-16 The United States Of America As Represented By The Secretary Of The Air Force Method for three-dimensional, inhomogeneity localization in turbid medium using diffuse photon density waves
US7006676B1 (en) 2000-01-21 2006-02-28 Medical Optical Imaging, Inc. Method and apparatus for detecting an abnormality within a host medium utilizing frequency-swept modulation diffusion tomography
CA2348150C (en) 2000-05-25 2007-03-13 Esam M.A. Hussein Non-rotating x-ray system for three-dimensional, three-parameter imaging
US20020033454A1 (en) * 2000-08-04 2002-03-21 Xuefeng Cheng Optical imaging system with direct image construction
US7383076B2 (en) * 2000-11-27 2008-06-03 The General Hospital Corporation Fluorescence-mediated molecular tomography
IL141135A0 (en) * 2001-01-28 2002-02-10 Israel Atomic Energy Comm Method for imaging in a turbid medium
US20020186893A1 (en) * 2001-04-06 2002-12-12 Marmarelis Vasilis Z. Nonlinear processing for mitigation of diffraction effects
US6930683B2 (en) * 2002-04-19 2005-08-16 Hewlett-Packard Development Company, L.P. Three-dimensional reconstruction method utilizing reprojective optimization
EP1514093B1 (en) 2002-06-04 2021-04-28 Visen Medical, Inc. Imaging volumes with arbitrary geometries in non-contact tomography
US7302376B2 (en) * 2002-08-15 2007-11-27 International Business Machines Corporation Device modeling for proximity effects
US7647091B2 (en) * 2003-02-05 2010-01-12 The General Hospital Corporation Method and system for free space optical tomography of diffuse media
EP1654531A1 (en) * 2003-06-20 2006-05-10 The Texas A &amp; M University System Method and system for near-infrared fluorescence contrast-enhanced imaging with area illumination and area detection
CA2513990C (en) * 2004-08-27 2010-09-14 Paul Jacob Arsenault X-ray scatter image reconstruction by balancing of discrepancies between detector responses, and apparatus therefor
JP2009511124A (en) * 2005-10-06 2009-03-19 イメージング・サイエンシィズ・インターナショナル・エルエルシー Scatter correction
DK2156370T3 (en) 2007-05-14 2012-01-23 Historx Inc Compartment separation by pixel characterization using image data clustering
EP2162728B1 (en) * 2007-06-15 2016-07-27 Novartis AG Microscope system and method for obtaining standardized sample data
CA2604317C (en) 2007-08-06 2017-02-28 Historx, Inc. Methods and system for validating sample images for quantitative immunoassays
CA2596204C (en) 2007-08-07 2019-02-26 Historx, Inc. Method and system for determining an optimal dilution of a reagent
US8454512B2 (en) 2007-10-25 2013-06-04 Washington University Confocal photoacoustic microscopy with optical lateral resolution
US8243071B2 (en) * 2008-02-29 2012-08-14 Microsoft Corporation Modeling and rendering of heterogeneous translucent materials using the diffusion equation
EP2335221B8 (en) 2008-09-16 2016-05-25 Novartis AG Reproducible quantification of biomarker expression
US8416421B2 (en) * 2008-10-01 2013-04-09 Washington University Optical coherence computed tomography
WO2010080991A2 (en) 2009-01-09 2010-07-15 Washington University In St. Louis Miniaturized photoacoustic imaging apparatus including a rotatable reflector
WO2011127428A2 (en) 2010-04-09 2011-10-13 Washington University Quantification of optical absorption coefficients using acoustic spectra in photoacoustic tomography
JP5813994B2 (en) * 2010-06-03 2015-11-17 株式会社東芝 Medical image diagnostic apparatus and image reconstruction method
US8997572B2 (en) 2011-02-11 2015-04-07 Washington University Multi-focus optical-resolution photoacoustic microscopy with ultrasonic array detection
US10216866B2 (en) * 2011-02-25 2019-02-26 Smiths Heimann Gmbh Image reconstruction based on parametric models
US8416914B2 (en) * 2011-07-08 2013-04-09 General Electric Company System and method of iterative image reconstruction for computed tomography
WO2013070923A1 (en) * 2011-11-08 2013-05-16 The Trustees Of Columbia University In The City Of New York Tomographic imaging methods, devices, and systems
US8989342B2 (en) * 2012-04-18 2015-03-24 The Boeing Company Methods and systems for volumetric reconstruction using radiography
JP6212312B2 (en) * 2012-08-13 2017-10-11 パナソニック株式会社 Object interior estimation apparatus and method
US11020006B2 (en) 2012-10-18 2021-06-01 California Institute Of Technology Transcranial photoacoustic/thermoacoustic tomography brain imaging informed by adjunct image data
US9091628B2 (en) 2012-12-21 2015-07-28 L-3 Communications Security And Detection Systems, Inc. 3D mapping with two orthogonal imaging views
WO2015077355A1 (en) 2013-11-19 2015-05-28 Washington University Systems and methods of grueneisen-relaxation photoacoustic microscopy and photoacoustic wavefront shaping
WO2018209046A1 (en) 2017-05-10 2018-11-15 Washington University Snapshot photoacoustic photography using an ergodic relay
WO2020037082A1 (en) 2018-08-14 2020-02-20 California Institute Of Technology Multifocal photoacoustic microscopy through an ergodic relay
US11592652B2 (en) 2018-09-04 2023-02-28 California Institute Of Technology Enhanced-resolution infrared photoacoustic microscopy and spectroscopy
US11369280B2 (en) 2019-03-01 2022-06-28 California Institute Of Technology Velocity-matched ultrasonic tagging in photoacoustic flowgraphy
CN110196239B (en) * 2019-06-12 2020-09-29 中国科学院南京地理与湖泊研究所 Spectrum remote sensing inversion method for phytoplankton absorption coefficient of turbid water body

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4480332A (en) * 1981-05-23 1984-10-30 U.S. Philips Corporation Apparatus for examination by scattered radiation
US4768214A (en) * 1985-01-16 1988-08-30 American Science And Engineering, Inc. Imaging
US4850002A (en) * 1986-09-09 1989-07-18 U.S. Philips Corporation Two-dimensional compton profile imaging method
US4870670A (en) * 1987-10-19 1989-09-26 Heimann Gmbh X-ray scanner with secondary radiation detector
US4896342A (en) * 1987-04-01 1990-01-23 U.S. Philips Corporation X-ray apparatus
US4945239A (en) * 1989-03-29 1990-07-31 Center For Innovative Technology Early detection of breast cancer using transillumination
US4948974A (en) * 1984-06-25 1990-08-14 Nelson Robert S High resolution imaging apparatus and method for approximating scattering effects
US4956856A (en) * 1987-10-05 1990-09-11 U.S. Philips Corporation Arrangement for examining a body comprising a radiation source

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB1283915A (en) * 1968-08-23 1972-08-02 Emi Ltd A method of and apparatus for examination of a body by radiation such as x or gamma radiation
US4422146A (en) * 1979-11-22 1983-12-20 The President Of Tokyo Institute Of Technology Reconstruction method of a computed tomographic image from a few X-ray projections
US4337799A (en) * 1980-09-22 1982-07-06 Hoover C Roger Thread protector

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4480332A (en) * 1981-05-23 1984-10-30 U.S. Philips Corporation Apparatus for examination by scattered radiation
US4948974A (en) * 1984-06-25 1990-08-14 Nelson Robert S High resolution imaging apparatus and method for approximating scattering effects
US4768214A (en) * 1985-01-16 1988-08-30 American Science And Engineering, Inc. Imaging
US4850002A (en) * 1986-09-09 1989-07-18 U.S. Philips Corporation Two-dimensional compton profile imaging method
US4896342A (en) * 1987-04-01 1990-01-23 U.S. Philips Corporation X-ray apparatus
US4956856A (en) * 1987-10-05 1990-09-11 U.S. Philips Corporation Arrangement for examining a body comprising a radiation source
US4870670A (en) * 1987-10-19 1989-09-26 Heimann Gmbh X-ray scanner with secondary radiation detector
US4945239A (en) * 1989-03-29 1990-07-31 Center For Innovative Technology Early detection of breast cancer using transillumination

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0627698A1 (en) * 1993-06-03 1994-12-07 Hamamatsu Photonics K.K. Optical computer tomograph
EP0633527A1 (en) * 1993-07-09 1995-01-11 Hamamatsu Photonics K.K. Optical computer tomograph
EP0692708A3 (en) * 1994-07-14 1996-10-30 Hitachi Ltd Imaging method for spatial distributions of absorber concentrations
US5678556A (en) * 1994-07-14 1997-10-21 Hitachi, Ltd. Imaging method for spatial distributions of absorber concentrations
US5719398A (en) * 1995-09-11 1998-02-17 U.S. Philips Corporation Method and device for imaging the interior of a turbid medium

Also Published As

Publication number Publication date
US5070455A (en) 1991-12-03
JPH05502393A (en) 1993-04-28

Similar Documents

Publication Publication Date Title
US5070455A (en) Imaging system and method using scattered and diffused radiation
Politte et al. Corrections for accidental coincidences and attenuation in maximum-likelihood image reconstruction for positron-emission tomography
DE69918596T2 (en) Computer tomograph with iterative reconstruction
EP0610318B1 (en) Method and apparatus for image processing
Wernecke Maximum entropy image reconstruction
US6339652B1 (en) Source-assisted attenuation correction for emission computed tomography
US6978039B2 (en) Method and system for segmentation of medical images
Skilling et al. Algorithms and applications
KR102548577B1 (en) Medical imaging device with a fixed X-ray detector array and a fixed X-ray emitter array forming a digital 3D image
US20040236549A1 (en) Method and apparatus for rapidly evaluating digital data processing parameters
JP2004033749A (en) Semiautomatic segmental algorithm of pet (positron emission tomograph) tumor image
US5414803A (en) Method utilizing frequency domain representations for generating two-dimensional views of three-dimensional objects
US20080123961A1 (en) Scalable method for rapidly detecting potential ground vehicle under cover using visualization of total occlusion footprint in point cloud population
JP2003532873A (en) Optical computed tomography in opaque media
Miller et al. 3-D maximum a posteriori estimation for single photon emission computed tomography on massively-parallel computers
Grunbaum et al. Diffuse tomography
US5758653A (en) Simultaneous absorption and diffusion imaging system and method using direct reconstruction of scattered radiation
Fleming Satellite remote sensing by the technique of computed tomography
Girard Optimal regularized reconstruction in computerized tomography
US6269323B1 (en) Support vector method for function estimation
US5746211A (en) Absorption imaging system and method using direct reconstruction of scattered radiation
Pawitan et al. Data-dependent bandwidth selection for emission computed tomography reconstruction
Singh et al. Noise propagation in electronically collimated single photon imaging
Wang et al. Multiobjective neural network for image reconstruction
Kent et al. Some suggestions for transmission tomography based on the EM algorithm

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): CA JP KR

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IT LU NL SE

NENP Non-entry into the national phase

Ref country code: CA