WO2010062884A1 - Image processing for curvature correction - Google Patents

Image processing for curvature correction Download PDF

Info

Publication number
WO2010062884A1
WO2010062884A1 PCT/US2009/065660 US2009065660W WO2010062884A1 WO 2010062884 A1 WO2010062884 A1 WO 2010062884A1 US 2009065660 W US2009065660 W US 2009065660W WO 2010062884 A1 WO2010062884 A1 WO 2010062884A1
Authority
WO
WIPO (PCT)
Prior art keywords
marks
image
unread
sector
locations
Prior art date
Application number
PCT/US2009/065660
Other languages
French (fr)
Inventor
Frank J. Metayer
Carlos Cervantes Toral
Duncan D. Xue
Original Assignee
Gtech Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Gtech Corporation filed Critical Gtech Corporation
Priority to AU2009319844A priority Critical patent/AU2009319844A1/en
Priority to EP09796868A priority patent/EP2366171A1/en
Priority to CA2744698A priority patent/CA2744698A1/en
Priority to GB1110263A priority patent/GB2477707A/en
Priority to JP2011538665A priority patent/JP2012510235A/en
Priority to CN2009801529378A priority patent/CN102265303A/en
Publication of WO2010062884A1 publication Critical patent/WO2010062884A1/en
Priority to IL213149A priority patent/IL213149A0/en

Links

Classifications

    • G06T3/14
    • G06T3/06
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10712Fixed beam scanning
    • G06K7/10722Photodetector array or CCD scanning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T1/00General purpose image data processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/80
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07FCOIN-FREED OR LIKE APPARATUS
    • G07F17/00Coin-freed apparatus for hiring articles; Coin-freed facilities or services
    • G07F17/32Coin-freed apparatus for hiring articles; Coin-freed facilities or services for games, toys, sports, or amusements
    • G07F17/3286Type of games
    • G07F17/3288Betting, e.g. on live events, bookmaking
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07FCOIN-FREED OR LIKE APPARATUS
    • G07F17/00Coin-freed apparatus for hiring articles; Coin-freed facilities or services
    • G07F17/32Coin-freed apparatus for hiring articles; Coin-freed facilities or services for games, toys, sports, or amusements
    • G07F17/3286Type of games
    • G07F17/329Regular and instant lottery, e.g. electronic scratch cards

Definitions

  • the present invention relates to an optical reader system employing a camera to capture a digital data image of an unread form and a computer system for processing the digital data to correct for distortions and derive and interpret information from the digital data.
  • camera includes any photo-sensitive device that captures a scene in a photo-sensitive media that may be downloaded or read-out by a computer processing system that stores in memory the digital data representing the scene. Lenses, filters, apertures, etc., and a photo-sensitive surface, array, or arrays for holding the scene image may be included in the term "camera.”
  • Forms are collectively defined herein as "forms.”
  • Forms y often, have man-made marks at locations indicating a specific intent of the human. Correctly identifying a form and reading or processing the printed and man-made markings are not simple tasks.
  • Some of these tasks include: detecting the presence of the form, determining that the form is still or motion-less, identifying the existence of marks on the form, locating the marks on the form; and, then interpreting the meaning of the marks. For each task, there must be enough light to illuminate the form with enough contrast with respect to the background to detect and identify those marks.
  • printed markings may be detected and interpreted, or a human may indicate the form type. The printed markings will normally include logos or other special marks used to identify the form type.
  • alignment marks may be printed and used by the processing system to accurately identify locations on the form. Then, the markings on the form may be accurately located and interpreted for their meaning.
  • the processing and identifying of marks on the form begins with a camera that captures an image of the form.
  • the captured image may be downloaded, stored and analyzed by a computing system. Analyzing the image may be accomplished by an "application” that may be software running in a computer system, firmware embedded in a hardware framework, a hardware state machine, or a combination thereof.
  • Distortions are a major cause for errors in reading or rejecting a form. Distortions manifest, at least, as location and resolution errors.
  • model images of known form types such as, for example, printed documents, play slips, lottery scratch tickets or lottery instant tickets are stored on a computer system.
  • information is stored on the computer system that includes known logo or other form identifying marks and alignment marks that partition each model image into areas or sectors that are defined by at least three non-colinear alignment marks.
  • the locations, attributes (see below) and the corresponding meanings of alignment marks, logos, other such identifying marks, and the locations of possible hand drawn marks and their meanings are also stored with each model image.
  • a coordinate system such as, for example, Cartesian, polar, etc., describes locations on a model image and on any captured images of unread forms.
  • x and y may define locations on the model image, and x 1 and y 1 may be used for locations on a unread image of a form.
  • a transfer function may be applied to an apparent location of any mark within a sector of the unread image to yield a corrected location. The apparent location may then be compared to the corresponding location on the model image and that location may determine the mark's meaning.
  • AU the sections having marks therein may be processed in a like manner.
  • Transfer function is a phrase used herein that may be termed by others as
  • Attributes may refer to shapes, sizes, the center of mass of the mark, etc.
  • the location and the attributes of a mark may define an alignment mark or a logo.
  • the type of form may be determined by a logo, or some other printed mark or marks, or may be manually made by an agent. For example, a pencil mark at a specific location on the form may indicate a selection by a human completing a form such as when playing a game.
  • An alignment mark on an unread image is easily detected and recognized by a computer system.
  • the mark found at a specific location, has attributes (large, round, square, oblong, long linear or any shape) that accommodate its detection as an alignment mark
  • logos are marks similar to alignment marks in their ease of detection and identification, and, as mentioned above, may be used to identify the type of form.
  • Partitioning a form into sectors allows for correction of curvature distortion by sizing the sectors. Smaller sectors may be relatively flat where curves in such a sector do not manifest to cause errors in reading marks in the sector. Where a sector is defined by three non-colinear points, the following equations may represent a transfer function that corrects scaling, translation and rotation of marks within that sector:
  • x' A'x + B'y + C Eq. 1
  • y 1 D'x + E'y + F 1 Eq. 2.
  • Equations 1 and 2 may be used to translate any location (x, y) in the model image to the corresponding coordinate (x 1 , y') in the unread image of the form.
  • a sector is defined by four non-colinear points that define a four sided sector area
  • the following equations may represent a transfer function that corrects for scaling, translation, rotation and for perspective issues associated with a camera capturing the unread sector of a form:
  • Equations 3 and 4 may be used to translate any location (x, y) in the model image to the corresponding coordinate (x 1 , y') in the unread image of the form.
  • the coefficients of the above transfer function equations are found by comparing the actual x 1 , y' locations of specific marks in the unread image to the corresponding x, y locations in the model image.
  • at least three non- colinear known alignment marks may be processed, however, four or more marks may be used to advantage.
  • a system of equations result where the coefficients may be found by use of matrix algebra techniques as illustrated below.
  • eight alignment marks are used to define a sector. According to this illustration, when eight alignment marks are used to define the sector, the following biquadratic transfer function equation (Equation 7) may be applied and correct for scaling, translation, rotation and perspective issues associated with a camera capturing the unread sector of a form:
  • the same sectoring may be applied to the stored unread image and the transfer functions may be applied to the locations of the marks within every sector on the unread image. This results in corrected locations for the marks, if any, within each sector. The corrected locations of these other marks then may be compared to corresponding locations on the model image. The locations on the model image have stored meanings that determine the meanings of those other marks whose corrected locations match those of the model image.
  • the transfer function may be applied to areas immediately outside that area
  • a sector includes other alignment marks or marks of known locations, those other marks may be used to calculate an error.
  • the locations of those other marks on the stored unread image may be detected and their locations may be corrected by application of the transfer function.
  • the corrected location may be compared to the actual location from the model image and an error may be calculated.
  • An error threshold may be established, and if the error in any sector exceeds the threshold, the form may be rejected.
  • other, closer alignment marks may be used to form yet another sector and another transfer function may be applied. This process may continue until there are no smaller sectors with acceptable errors, wherein the form may be rejected. If an acceptable error is formed for every sector in the form where a mark exists, the marked form may be accepted and may then be processed by the system.
  • FIG. 1 is a system block diagram of a system embodying the present invention
  • FIGs. 2A, 2B and 2C are a series of illustrations of curvature problems
  • FIG. 3 is a flow chart illustrating a process of the present invention
  • FIG. 4 is a layout of a form with markings
  • FIGs. 5 A, 6A and 7A are examples of examples of forms that may be read using one or more processes of the present invention.
  • FIGs. 5B, 6B, and 7B illustrate the partitioning of the forms shown in Figs. 5 A, 6 A, and 7 A, respectively, into sectors in accord with one or more of the processes described in the present patent.
  • FIG. 1 illustrates a form 2 being illuminated by a light source 4 with the reflected light 6 from the form 2 being received by a camera 18.
  • the form may, for example, represent a printed document, a play slip, a lottery scratch ticket or a lottery instant ticket.
  • the form 2 may be located on a platen 5 positioned a few inches to a foot or more away from the camera 18. The arrangement allows the entire form 2 to be viewed by the camera 18, although sections of the camera scene, and, thus, the form 2 may be processed in sequence.
  • An image is impressed on the camera's photo-sensitive surface and may be downloaded (scanned or read-out) by camera electronics 8 to produce a video signal 10, that is digitized and stored as pixels or pixel data, and where a processor 14 in a computer processing system 12, preferably a micro-processor, operates on the pixel data.
  • the computer processing system 12 includes memory 18 and I/O device drivers 16 for handling, at least, displays, keyboards, buttons, printers, and communications.
  • the computer processing system 12, in turn, may be connected with a computer network 17.
  • Memory 18 may include, at least, one or more image buffers, other buffers, cache, etc.
  • An operating system and software applications may be stored in the memory 18.
  • a processing application 13, discussed below, includes the processes for correcting or compensating for the curvature or distortion of a bent form.
  • an external, removable flash memory 19 may contain proprietary software, that is uploaded to the computer processor system 12. In this embodiment, removing the flash memory 19 removes the proprietary software from the computer processor system 12.
  • the computer network 17 may be a public network, such as the Internet, or a private network, etc., but in any case, information, such as, for example, financial information, may be encrypted to prevent outside interference and protection.
  • information may be deriyed from the form and sent via the network 17 to a central controller 15.
  • the central controller 15 verifies the financial information and authorizes agents to act accordingly.
  • FIGs. 2A and 2B illustrate some issues that are presented when a bent, rolled, folded or crumpled form 2 is placed before and viewed by a camera 18.
  • a form 2 with a single bend 20 is placed on a platen 5 and illuminated by light 4 from a light source such as an LED array 22.
  • a square mark 24 lies on the bent portion of the form 2, and light 26 from the light source is reflected from the square 24 to the camera 18
  • the beam of light 26 subtends an angle ⁇ , which is the same angle ⁇ that the light would strike the platen 5 at a location 24' in the absence of the form 2.
  • FIG. 2B shows a side elevation view of the same illustration set forth in FIG. 2A, where the beam of light 26 intersects the bent form 2 at the square 24 and intersects the platen 5 at 24'.
  • the length of the side view of the square 24 is much shorter than the length of the side view of the square 24'.
  • a single dot A is on the bent portion of the form 2.
  • a small beam of light 26 is reflected from A to the camera 18.
  • dot A will be smaller in size to the projection dot A'.
  • the bent portion is flattened 28 to the platen 5
  • the dot A is located at location A".
  • the image analyzer will process the mark A as if it were located at location A' whereas the true position is at location A".
  • the bend 20 in the form 2 produces at least a size error and a position error. If the form 2 is curved, curled or has multiple bends the resulting errors will be that much more difficult to accommodate.
  • FIG. 3 illustrates a method for processing pixel data of a known form according to the present invention.
  • pixel data of an unread image is compared to a model image of the form stored in the processing system memory 18 (FIG. 1).
  • the model image is used as a template that contains the locations of all the alignment marks.
  • the model image form is divided into sectors where the number of alignment marks define each sector and include a transfer function for each sector.
  • the locations of the alignments marks in the stored captured image and the model image determine the coefficients of the transfer functions.
  • the transfer functions are then applied to the locations of marks in the corresponding sectors in the unread images to produce corrected locations, and the meanings of these marks are determined by reference to the corresponding locations in the model image.
  • Meanings of locations of marks on the model image are stored.
  • the process illustrated in FIG. 3 starts with a captured unread image of a known form type that includes marks such as alignment dots or marks 31.
  • a stored model image that corresponds with the unread image of the particular form is retrieved 32.
  • the locations of the alignment marks on the captured image are found 33, and, using the locations of corresponding alignment marks on the model image, the coefficients of a transfer function are found 34.
  • a transfer function has been determined for a sector, that function may be applied to other alignment marks in or near that sector.
  • an error 35 may be calculated.
  • An error threshold may be established for an error that can be tolerated, and if the error is within the threshold, the form may be processed further as discussed herein. If the error is too large other closer alignment marks may be used, as discussed herein, until an acceptable error is found. If an unacceptable error remains, the form may be rejected 46.
  • An acceptable error may be one that is smaller than the position error that an analyzing system assigns to an acceptable known location on the form.
  • the area where a mark may be made could be a square of 0.1 inches on a side.
  • a position error of 0.05 inches may be an acceptable error.
  • the transfer function for each sector is applied to all the marks within that sector 42, producing corrected locations for all the marks on the captured image 44.
  • the meanings of all the marks, as mentioned above, are found 48 by reference to corresponding locations on the model image.
  • a form may be divided into triangular and rectangular sectors, but any figure encompassing an area may be used.
  • acceptable errors may be determined by running sample or test forms until the rejection rates of the forms is acceptable. In some applications, acceptable error may be simply estimated.
  • FIG. 4 illustrates a top view of a form 50 that is laid flat. Alignment marks 54a, 54b, 54c, and 54d, and 54a', 54b 1 , 54c', and 54d' are arranged as shown.
  • the calculated distance in a stored captured image between alignment marks 54a and 54b may be only 80% of what that distance is on the model image. Using that factor, the distance on the captured image from mark C to mark 54a may be corrected by multiplying that distance by 1/0.8, or by 1.25.
  • Mark C would then be interpreted as a mark at that location on the model image, and the significance of that location on the model will be assigned to mark C.
  • Mark C could be a logo, another alignment mark, or a man-made mark. Note that if the alignment marks were closer together the actual distance error will be reduced accordingly. That is, if a distance of five inches is corrected by 1%, the correction is 0.05 inch, but if the distance is 2.5 inches the same correction factor would be only 0.025 inches. Thus, the closer the alignment marks the smaller the error.
  • Detection of marks of any kind on a form entails a contrast that allows the edges of the marks, and then characteristics or attributes of the marks to be found.
  • every form type will have a logo, that is quite different in attributes so that it is easily found and recognized, if one is not found, the form type may be keyed in by an agent or the form rejected.
  • alignment marks are distinctive to ease detection.
  • analyzing four alignment marks, 54a, 54b, 54c, and 54d, at the four corners of the stored captured image and a model image form may produce a transfer function that when applied results in an acceptable error. If so, the other marks in the stored captured image may be processed by that transfer function. Such an example would quicken the processing of that form.
  • additional alignment marks may be employed. If any errors are deemed too large, in some applications, additional alignment marks (for example, 54a', 54b 1 , 54c', 54d') may be employed. If these eight alignment marks are used, the form may be divided into three sectors, 52, 53, and 55. A transfer function may be developed for each sector and any marks found within each sector would be corrected using the corresponding transfer function. Alignment marks, moreover, may be placed, where acceptable, within the central portion D of the form. As more sections are used, the processing of the form may take longer. Note, however, that if no other marks are in a sector, that sector need not be processed.
  • FIGs. 5 A, 6A and 7A illustrate three exemplary forms.
  • FIG. 5A has eleven alignment marks 60 distributed around the outer periphery of the form. Each alignment mark 60 has a known shape for that form type.
  • FIG. 6A depicts another form type with alignment marks 70 distributed also within the center portion of the form. This form type has ten peripheral alignment marks 72 along with the two 70 central alignment marks.
  • FIG. 7A is yet another form type. Here, the alignment marks are positioned similarly to those in FIG. 6A, but the center marks 76 and the periphery marks 74 are thinner and not as bold in comparison to those in FIG. 6 A. In FIG. 7A, the alignment marks are less obvious and intrusive to a casual viewer.
  • FIGs. 5B, 6B, and 7B illustrate sectoring that may be applied to the corresponding form types shown in Figs. 5A, 6A and 7A.
  • FIG. 5B there are five triangular sectors
  • FIG. 5B has one rectangular sector 87 and another four-sided sector 86. These four-sided sectors would have a four point transfer function applied.
  • FIGs. 6B and 7B are shown with six rectangular or square sectors 90 and 92, respectively. Four point transfer functions would be developed for each of these sectors, The Transfer Functions:
  • each model image of a form may be partitioned into a number of sectors, some defined by three alignment marks, others by four alignment marks and yet others may be defined by more than four such marks.
  • Eqs. 1 and 2 may be applied to translate any location (x, y) on the model image to the corresponding coordinate (x 1 , y 1 ) on the captured image.
  • Eqs. 1 and 2 may be arranged into matrix form as follows:
  • a transfer function that may apply to the locations within the sector may be defined as follows:
  • Eq. 4 Once the eight coefficients (A, B, C, D, E, F, G, and H) are derived for a specific captured image, Eqs. 3 and 4 can be used to translate any location (x, y) on the model image to the corresponding coordinate (x 1 , y 1 ) in the captured image.
  • Eqs. 3 and 4 may be rearranged in more matrix-like form as follows:
  • Equation 7 When more than four alignment marks are used to define a sector, a transfer function that may apply to the locations within the sector may be defined by the following biquadratic equation, which represents a bi-quadratic transformation: Equation 7:
  • a and B" are coefficients matrices of the transformation and can be expressed as two 1 x 8 matrices respectively:
  • U is an 1 x 8 matrix, its transpose matrix, U ⁇ , is an 8 x 1 matrix.
  • U ⁇ instead of U is defined below for clarity:
  • This eight-point transfer function needs eight reference points to determine the coefficients matrices A" and B". Assume one has eight reference points in the template or model image:
  • Equation 8 Equation 8
  • Equation 9 If matrix Q is non-singular, one can solve the coefficient matrix A and B from Eq. 2 and obtain Equation 9:
  • x' j and y' j are known coordinates of the pre-selected eight reference points in the captured space.
  • Q -1 is also known since it depends only on the coordinates ⁇ X j ,y j ) of the pre-selected eight reference points in the template or model image.
  • U n (x, y) can be calculated. Therefore, for any point ( x , y ) in the template, the corresponding point ( x' , y' ) of the captured space can be obtained from the above two equations.
  • Q -1 can be computed offline to speed up the transformation. That means, we do not have to compute Q -1 for every pair of ( x , y ) to (x' , y' ) transformation. It can be computed only once after the eight reference points in the template are selected.

Abstract

A system and process for reading forms using a curvature distortion correction scheme is disclosed. A model image of a known form is partitioned into sectors defined by at least three alignment marks, and a transfer or correction function for each sector is generated by comparing the locations of alignment marks on a captured unread image of a form to those on a model image. The transfer function is then applied to the other marks within the sector to calculate a correct location of the marks. Each sector is sized to reasonably ensure curvature distortions are handled such that all the marks on a form are properly found and interpreted. Errors may be checked and, if they persist, smaller sectors may be defined and employed, if available.

Description

IMAGE PROCESSING FOR CURVATURE CORRECTION
RELATED APPLICATIONS
The present invention is a continuation-in-part of U.S. Patent Application No. 11/376,052, filed March 14, 2006 and entitled SYSTEM AND METHOD FOR PROCESSING A FORM, that claims priority from U.S. Provisional Patent Application Serial No. 60/661,698, filed on March 14, 2005. These patent applications are of common ownership with the present invention, and they are incorporated herein by reference.
BACKGROUND OF THE INVENTION Field of the Invention
The present invention relates to an optical reader system employing a camera to capture a digital data image of an unread form and a computer system for processing the digital data to correct for distortions and derive and interpret information from the digital data. Herein, "camera" includes any photo-sensitive device that captures a scene in a photo-sensitive media that may be downloaded or read-out by a computer processing system that stores in memory the digital data representing the scene. Lenses, filters, apertures, etc., and a photo-sensitive surface, array, or arrays for holding the scene image may be included in the term "camera."
Background Information Printed documents, play slips, lottery scratch tickets, lottery instant tickets and the
- like are collectively defined herein as "forms." Formsy often, have man-made marks at locations indicating a specific intent of the human. Correctly identifying a form and reading or processing the printed and man-made markings are not simple tasks.
Some of these tasks include: detecting the presence of the form, determining that the form is still or motion-less, identifying the existence of marks on the form, locating the marks on the form; and, then interpreting the meaning of the marks. For each task, there must be enough light to illuminate the form with enough contrast with respect to the background to detect and identify those marks. In order to identify a form, printed markings may be detected and interpreted, or a human may indicate the form type. The printed markings will normally include logos or other special marks used to identify the form type. In addition, alignment marks may be printed and used by the processing system to accurately identify locations on the form. Then, the markings on the form may be accurately located and interpreted for their meaning.
The processing and identifying of marks on the form begins with a camera that captures an image of the form. The captured image may be downloaded, stored and analyzed by a computing system. Analyzing the image may be accomplished by an "application" that may be software running in a computer system, firmware embedded in a hardware framework, a hardware state machine, or a combination thereof.
One important and common problem when reading forms occurs when the form is bent, crumpled, rolled or folded, thereby creating distortions. Distortions are a major cause for errors in reading or rejecting a form. Distortions manifest, at least, as location and resolution errors.
Some prior art readers have drive mechanisms that flatten and position a form into an enclosed chamber. Such drive mechanisms are expensive and require routine service and technical support to maintain them, all leading to additional costs and expense in ownership. Prior art attempts at compensating for distortions have been ineffectual and a convenient, reliable, comparatively inexpensive easy to use reader that corrects, reduces and/or compensates for such distortions would benefit the industry.
SUMMARY OF THE INVENTION
The present invention is directed to an optical reader that is capable of reading a wide assortment of forms and, more particularly, capable of reading forms that have been bent, rolled, crumpled, or folded, by using a system and a process for distortion correction. According to at least one embodiment of the present invention, model images of known form types such as, for example, printed documents, play slips, lottery scratch tickets or lottery instant tickets are stored on a computer system. Along with the model images, information is stored on the computer system that includes known logo or other form identifying marks and alignment marks that partition each model image into areas or sectors that are defined by at least three non-colinear alignment marks. The locations, attributes (see below) and the corresponding meanings of alignment marks, logos, other such identifying marks, and the locations of possible hand drawn marks and their meanings are also stored with each model image.
A coordinate system such as, for example, Cartesian, polar, etc., describes locations on a model image and on any captured images of unread forms. In a Cartesian coordinate system, for example, x and y may define locations on the model image, and x1 and y1 may be used for locations on a unread image of a form. A transfer function may be applied to an apparent location of any mark within a sector of the unread image to yield a corrected location. The apparent location may then be compared to the corresponding location on the model image and that location may determine the mark's meaning. AU the sections having marks therein may be processed in a like manner. "Transfer function" is a phrase used herein that may be termed by others as
"translation function," or "equation," or "correction function," or "factor," or other such terms or combinations of terms as may be found in the art.
Attributes may refer to shapes, sizes, the center of mass of the mark, etc. The location and the attributes of a mark may define an alignment mark or a logo. The type of form may be determined by a logo, or some other printed mark or marks, or may be manually made by an agent. For example, a pencil mark at a specific location on the form may indicate a selection by a human completing a form such as when playing a game.
An alignment mark on an unread image is easily detected and recognized by a computer system. The mark, found at a specific location, has attributes (large, round, square, oblong, long linear or any shape) that accommodate its detection as an alignment mark Logos are marks similar to alignment marks in their ease of detection and identification, and, as mentioned above, may be used to identify the type of form.
Partitioning a form into sectors allows for correction of curvature distortion by sizing the sectors. Smaller sectors may be relatively flat where curves in such a sector do not manifest to cause errors in reading marks in the sector. Where a sector is defined by three non-colinear points, the following equations may represent a transfer function that corrects scaling, translation and rotation of marks within that sector:
x' = A'x + B'y + C Eq. 1, and y1 = D'x + E'y + F1 Eq. 2.
Once the six coefficients (A', B', C, D', E', and F') are derived for a specific captured image of the form, Equations 1 and 2 may be used to translate any location (x, y) in the model image to the corresponding coordinate (x1, y') in the unread image of the form.
Where a sector is defined by four non-colinear points that define a four sided sector area, the following equations may represent a transfer function that corrects for scaling, translation, rotation and for perspective issues associated with a camera capturing the unread sector of a form:
, Ax + By + C x — Eq. 3, and
Gx + Hy + l
, Dx + Ey + F
Eq. 4.
Gx + Hy + l
Once the eight coefficients (A1 R. C , D, E, F, G, unread image of the form, Equations 3 and 4 may be used to translate any location (x, y) in the model image to the corresponding coordinate (x1, y') in the unread image of the form.
The coefficients of the above transfer function equations are found by comparing the actual x1, y' locations of specific marks in the unread image to the corresponding x, y locations in the model image. In order to cover an area of the form, at least three non- colinear known alignment marks may be processed, however, four or more marks may be used to advantage. A system of equations result where the coefficients may be found by use of matrix algebra techniques as illustrated below. In the following illustration, eight alignment marks are used to define a sector. According to this illustration, when eight alignment marks are used to define the sector, the following biquadratic transfer function equation (Equation 7) may be applied and correct for scaling, translation, rotation and perspective issues associated with a camera capturing the unread sector of a form:
x1 = α, + Ct2X + a^y + aΛxy + a5x2 + a6y2 + a7xy2 + asx2y ≡ ∑anun(x,y) = UA"
«=1,8 γ = bx + b2x + b3y + b4xy + b5x2 + b6y2 + b7xy2 + bsx2y ≡ ∑bnu,,(x, y) = UB" n=l,8
After a form type is known and a transfer function is derived for every sector in the form, the same sectoring may be applied to the stored unread image and the transfer functions may be applied to the locations of the marks within every sector on the unread image. This results in corrected locations for the marks, if any, within each sector. The corrected locations of these other marks then may be compared to corresponding locations on the model image. The locations on the model image have stored meanings that determine the meanings of those other marks whose corrected locations match those of the model image.
In some applications, however the transfer function may be applied to areas immediately outside that area
If a sector includes other alignment marks or marks of known locations, those other marks may be used to calculate an error. The locations of those other marks on the stored unread image may be detected and their locations may be corrected by application of the transfer function. The corrected location may be compared to the actual location from the model image and an error may be calculated. An error threshold may be established, and if the error in any sector exceeds the threshold, the form may be rejected. In other applications, other, closer alignment marks may be used to form yet another sector and another transfer function may be applied. This process may continue until there are no smaller sectors with acceptable errors, wherein the form may be rejected. If an acceptable error is formed for every sector in the form where a mark exists, the marked form may be accepted and may then be processed by the system. It will be appreciated by those skilled in the art that although the following Detailed Description will proceed with reference being made to illustrative embodiments, the drawings, and the described methods of use, the present invention is not intended to be limited to these embodiments or methods of use. Rather, the present invention is of broad scope and is intended to be defined as only set forth in the accompanying claims.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention description below refers to the accompanying drawings, of which: FIG. 1 is a system block diagram of a system embodying the present invention;
FIGs. 2A, 2B and 2C are a series of illustrations of curvature problems;
FIG. 3 is a flow chart illustrating a process of the present invention;
FIG. 4 is a layout of a form with markings;
FIGs. 5 A, 6A and 7A are examples of examples of forms that may be read using one or more processes of the present invention; and
FIGs. 5B, 6B, and 7B illustrate the partitioning of the forms shown in Figs. 5 A, 6 A, and 7 A, respectively, into sectors in accord with one or more of the processes described in the present patent.
DETAILED DESCRIPTION OF AN ILLUSTRATIVE EMBODIMENT
FIG. 1 illustrates a form 2 being illuminated by a light source 4 with the reflected light 6 from the form 2 being received by a camera 18. The form may, for example, represent a printed document, a play slip, a lottery scratch ticket or a lottery instant ticket. The form 2 may be located on a platen 5 positioned a few inches to a foot or more away from the camera 18. The arrangement allows the entire form 2 to be viewed by the camera 18, although sections of the camera scene, and, thus, the form 2 may be processed in sequence. An image is impressed on the camera's photo-sensitive surface and may be downloaded (scanned or read-out) by camera electronics 8 to produce a video signal 10, that is digitized and stored as pixels or pixel data, and where a processor 14 in a computer processing system 12, preferably a micro-processor, operates on the pixel data. The computer processing system 12 includes memory 18 and I/O device drivers 16 for handling, at least, displays, keyboards, buttons, printers, and communications. The computer processing system 12, in turn, may be connected with a computer network 17. Memory 18 may include, at least, one or more image buffers, other buffers, cache, etc. An operating system and software applications may be stored in the memory 18. A processing application 13, discussed below, includes the processes for correcting or compensating for the curvature or distortion of a bent form. In one embodiment, an external, removable flash memory 19 may contain proprietary software, that is uploaded to the computer processor system 12. In this embodiment, removing the flash memory 19 removes the proprietary software from the computer processor system 12.
The computer network 17 may be a public network, such as the Internet, or a private network, etc., but in any case, information, such as, for example, financial information, may be encrypted to prevent outside interference and protection. In this example, information may be deriyed from the form and sent via the network 17 to a central controller 15. The central controller 15, in turn, verifies the financial information and authorizes agents to act accordingly.
FIGs. 2A and 2B (not to scale) illustrate some issues that are presented when a bent, rolled, folded or crumpled form 2 is placed before and viewed by a camera 18. In this illustration, a form 2 with a single bend 20 is placed on a platen 5 and illuminated by light 4 from a light source such as an LED array 22. In this illustration, a square mark 24 lies on the bent portion of the form 2, and light 26 from the light source is reflected from the square 24 to the camera 18 The beam of light 26 subtends an angle θ, which is the same angle θ that the light would strike the platen 5 at a location 24' in the absence of the form 2. The size of the square 24 on the form 2, however, is smaller than the square 24' on the platen 5 due to the spread of the light beam 26. The computer processor system 12 is ignorant of the bend 20 and processes the mark at location 24 as if it were at and as large as the mark at 24'. If the analysis of the camera image uses size as a factor in analyzing a mark, the bend will introduce a size error in the image. FIG. 2B shows a side elevation view of the same illustration set forth in FIG. 2A, where the beam of light 26 intersects the bent form 2 at the square 24 and intersects the platen 5 at 24'. The length of the side view of the square 24 is much shorter than the length of the side view of the square 24'. There are additional problems, however, as illustrated by the diagram in FIG. 2C.
In FIG. 2C, a single dot A is on the bent portion of the form 2. A small beam of light 26 is reflected from A to the camera 18. As above, dot A will be smaller in size to the projection dot A'. But, note that when the bent portion is flattened 28 to the platen 5, the dot A is located at location A". There is an offset 30, a position error, between the projections at A' and A". The image analyzer will process the mark A as if it were located at location A' whereas the true position is at location A". The bend 20 in the form 2 produces at least a size error and a position error. If the form 2 is curved, curled or has multiple bends the resulting errors will be that much more difficult to accommodate.
FIG. 3 illustrates a method for processing pixel data of a known form according to the present invention. According to this process, and in broad terms, pixel data of an unread image is compared to a model image of the form stored in the processing system memory 18 (FIG. 1). The model image is used as a template that contains the locations of all the alignment marks. The model image form is divided into sectors where the number of alignment marks define each sector and include a transfer function for each sector. The locations of the alignments marks in the stored captured image and the model image determine the coefficients of the transfer functions. The transfer functions are then applied to the locations of marks in the corresponding sectors in the unread images to produce corrected locations, and the meanings of these marks are determined by reference to the corresponding locations in the model image. Meanings of locations of marks on the model image are stored. To restate in more detail, the process illustrated in FIG. 3 starts with a captured unread image of a known form type that includes marks such as alignment dots or marks 31. A stored model image that corresponds with the unread image of the particular form is retrieved 32. The locations of the alignment marks on the captured image are found 33, and, using the locations of corresponding alignment marks on the model image, the coefficients of a transfer function are found 34. When a transfer function has been determined for a sector, that function may be applied to other alignment marks in or near that sector. When the comparison is made of the corrected locations of alignment marks in the captured image to the locations of those marks in the model image, an error 35 may be calculated. An error threshold may be established for an error that can be tolerated, and if the error is within the threshold, the form may be processed further as discussed herein. If the error is too large other closer alignment marks may be used, as discussed herein, until an acceptable error is found. If an unacceptable error remains, the form may be rejected 46.
An acceptable error may be one that is smaller than the position error that an analyzing system assigns to an acceptable known location on the form. For example, the area where a mark may be made could be a square of 0.1 inches on a side. A position error of 0.05 inches may be an acceptable error.
If the errors are all acceptable 40, the transfer function for each sector is applied to all the marks within that sector 42, producing corrected locations for all the marks on the captured image 44. The meanings of all the marks, as mentioned above, are found 48 by reference to corresponding locations on the model image.
If the error remains unacceptable, smaller sectors may be successively employed 45 with other alignment marks until a transfer function results in acceptable error for marks in every section of the form. If acceptable errors are not found, the form may be rejected 46.
Illustratively, a form may be divided into triangular and rectangular sectors, but any figure encompassing an area may be used.
In other applications, acceptable errors may be determined by running sample or test forms until the rejection rates of the forms is acceptable. In some applications, acceptable error may be simply estimated.
The following describes a simplistic example illustrating an application of a linear correcting process. FIG. 4 illustrates a top view of a form 50 that is laid flat. Alignment marks 54a, 54b, 54c, and 54d, and 54a', 54b1, 54c', and 54d' are arranged as shown. For example, for an illustrative rudimentary linear correction factor, the calculated distance in a stored captured image between alignment marks 54a and 54b may be only 80% of what that distance is on the model image. Using that factor, the distance on the captured image from mark C to mark 54a may be corrected by multiplying that distance by 1/0.8, or by 1.25. Mark C would then be interpreted as a mark at that location on the model image, and the significance of that location on the model will be assigned to mark C. Mark C could be a logo, another alignment mark, or a man-made mark. Note that if the alignment marks were closer together the actual distance error will be reduced accordingly. That is, if a distance of five inches is corrected by 1%, the correction is 0.05 inch, but if the distance is 2.5 inches the same correction factor would be only 0.025 inches. Thus, the closer the alignment marks the smaller the error.
Detection of marks of any kind on a form entails a contrast that allows the edges of the marks, and then characteristics or attributes of the marks to be found. Typically, every form type will have a logo, that is quite different in attributes so that it is easily found and recognized, if one is not found, the form type may be keyed in by an agent or the form rejected. Similarly, alignment marks are distinctive to ease detection.
There are known techniques for detecting edges or marks and for calculating attributes of marks, and are not further discussed herein.
In the embodiment of FIG. 4, analyzing four alignment marks, 54a, 54b, 54c, and 54d, at the four corners of the stored captured image and a model image form may produce a transfer function that when applied results in an acceptable error. If so, the other marks in the stored captured image may be processed by that transfer function. Such an example would quicken the processing of that form.
If any errors are deemed too large, in some applications, additional alignment marks (for example, 54a', 54b1, 54c', 54d') may be employed. If these eight alignment marks are used, the form may be divided into three sectors, 52, 53, and 55. A transfer function may be developed for each sector and any marks found within each sector would be corrected using the corresponding transfer function. Alignment marks, moreover, may be placed, where acceptable, within the central portion D of the form. As more sections are used, the processing of the form may take longer. Note, however, that if no other marks are in a sector, that sector need not be processed.
FIGs. 5 A, 6A and 7A illustrate three exemplary forms. FIG. 5A has eleven alignment marks 60 distributed around the outer periphery of the form. Each alignment mark 60 has a known shape for that form type. FIG. 6A depicts another form type with alignment marks 70 distributed also within the center portion of the form. This form type has ten peripheral alignment marks 72 along with the two 70 central alignment marks. FIG. 7A is yet another form type. Here, the alignment marks are positioned similarly to those in FIG. 6A, but the center marks 76 and the periphery marks 74 are thinner and not as bold in comparison to those in FIG. 6 A. In FIG. 7A, the alignment marks are less obvious and intrusive to a casual viewer.
FIGs. 5B, 6B, and 7B illustrate sectoring that may be applied to the corresponding form types shown in Figs. 5A, 6A and 7A. In FIG. 5B, there are five triangular sectors
81, 82, 83, 84, and 85, and each is formed by three points, and would have a three point transfer function applied as shown herein. FIG. 5B has one rectangular sector 87 and another four-sided sector 86. These four-sided sectors would have a four point transfer function applied.
FIGs. 6B and 7B are shown with six rectangular or square sectors 90 and 92, respectively. Four point transfer functions would be developed for each of these sectors, The Transfer Functions:
Restating from above, since a rectangular form, for example, is only a sheet of paper, it often will have distortions that might cause reading problems. Straight lines, etc. on the paper form will not necessarily be straight lines on the captured image. The template or model image of the form, however, will represent a flat, usually rectangle form with undistorted marks and lines. The present invention provides for non-linear transfer functions that compensate or correct for curvature distortions on the captured image.
As mentioned above, according to one or more embodiments of the present invention, each model image of a form may be partitioned into a number of sectors, some defined by three alignment marks, others by four alignment marks and yet others may be defined by more than four such marks.
THREEPOINTEOUATIONSAND SOLUTION
If a sector is defined by three alignment marks, a transfer function that applies to the locations within the sector may be as follows: x'=A'x + B'y + C Eq. 1, and y'=D'x + E'y + P Eq. 2.
Once the six coefficients (A, B, C, D, E, and F) are derived for a specific captured image, Eqs. 1 and 2 may be applied to translate any location (x, y) on the model image to the corresponding coordinate (x1, y1) on the captured image.
The following discussion illustrates one way to solve these equations. Eqs. 1 and 2 may be arranged into matrix form as follows:
To solve for A, B, and C:
Figure imgf000014_0004
To solve for D, E, and F:
Figure imgf000014_0005
Because of the simplicity of the matrix, Cramer's Rule can be used to solve this easily. For example:
Figure imgf000014_0003
FOUR POINTEOUATIONSAND SOLUTION
If a sector is defined by four alignment marks, a transfer function that may apply to the locations within the sector may be defined as follows:
Eq 3 and
Figure imgf000014_0001
y
Eq. 4.
Figure imgf000014_0002
Once the eight coefficients (A, B, C, D, E, F, G, and H) are derived for a specific captured image, Eqs. 3 and 4 can be used to translate any location (x, y) on the model image to the corresponding coordinate (x1, y1) in the captured image.
The following discussion illustrates one way to solve these equations. Eqs. 3 and 4 may be rearranged in more matrix-like form as follows:
A●x + B●y + C● 1 + D●O + E●O + F●O - G●x●x' - H●y●x1 = x1 Eq. 5, and
A●O + B●O + C●O + D●x + E●y + F1I - G●x● y' - H●y●y1 = y' Eq. 6.
Eqs. 5 and 6, then, may be solved as a system of equations using Gaussian
Elimination. According to Gaussian Elimination, assembly of the following 8 x 9 augmented matrix is required and upper triangulation and back substitution is applied:
)
)
)
)
)
)
)
Figure imgf000015_0001
BIQUADRATIC TRANSFORMATION APPLIED TO CURVATURE CORRECTION
When more than four alignment marks are used to define a sector, a transfer function that may apply to the locations within the sector may be defined by the following biquadratic equation, which represents a bi-quadratic transformation: Equation 7:
Figure imgf000016_0001
Where A" and B" are coefficients matrices of the transformation and can be expressed as two 1 x 8 matrices respectively:
Figure imgf000016_0002
And U is an 1 x 8 matrix, its transpose matrix, Uτ , is an 8 x 1 matrix. Uτ instead of U is defined below for clarity:
1
Figure imgf000016_0003
That means, un = un (x, y), with n = 1, 2, ..., 8, are functions of x and y.
This eight-point transfer function needs eight reference points to determine the coefficients matrices A" and B". Assume one has eight reference points in the template or model image:
Figure imgf000017_0001
And also eight corresponding points in the captured space:
Figure imgf000017_0002
Substituting these eight points to Eq. 1, one has (for j = 1, 2, ..., 8):
Figure imgf000017_0003
Or in matrix form, the above two lines become: Equation 8:
Figure imgf000017_0004
Where matrices A and B are the same as above, and matrices X', Y', Q are defined below:
Figure imgf000017_0005
In Eq. 8, X' , Y' , Q are all known, but A and B are unknowns to be solved. If matrix Q is singular, the above two linear systems given by Eq. 2 have no solution and one must select different 8 reference points than above.
If matrix Q is non-singular, one can solve the coefficient matrix A and B from Eq. 2 and obtain Equation 9:
A" = Q-1X' B" = Q-1Y'
Where Q-1 is the inverse of matrix Q . Substituting A and B Into Eq. 1 leads to
Equation 10: x'= UA" = UQ-1X'
/= UB" = UQ-1Y' Or equivalently, Equation 11 :
Figure imgf000018_0001
Here, x'j and y'j , with j = 1, 2, ..., 8, are known coordinates of the pre-selected eight reference points in the captured space. Q-1 is also known since it depends only on the coordinates {Xj ,yj ) of the pre-selected eight reference points in the template or model image. For any given point ( x , y ) in the template, Un (x, y) , with n = 1 , 2, ... , 8, can be calculated. Therefore, for any point ( x , y ) in the template, the corresponding point ( x' , y' ) of the captured space can be obtained from the above two equations. Actually Q-1 can be computed offline to speed up the transformation. That means, we do not have to compute Q-1 for every pair of ( x , y ) to (x' , y' ) transformation. It can be computed only once after the eight reference points in the template are selected.
It should be understood that above-described embodiments are being presented herein as examples and that many variations and alternatives thereof are possible. Accordingly, the present invention should be viewed broadly as being defined only as set forth in the hereinafter appended claims.
What is claimed is:

Claims

CLAIMS 1. A process for reading information on an unread form, the process comprising the steps of: storing an image of a model of the unread form in a computer system, the model image containing alignment marks; storing an image of the unread form in the computer system, the unread image containing alignment marks that correspond to the alignments marks on the model image; partitioning the model image into a group of sectors, wherein each sector is defined by the locations of at least three alignment marks; detecting the sectors on the unread image that correspond to the sectors on the model image by comparing the alignment marks on the model image to the corresponding alignment marks on the unread image; generating a transfer function for each sector of the unread image from the locations of corresponding alignments marks of the model and unread images; and applying each sector's transfer function to the locations of marks on the corresponding sectors of the unread image to generate corrected locations of the marks, thereby enabling the reading of the unread form.
2. The process of claim 1, further comprising the steps of: forming a coordinate system for each sector of the unread image, wherein the transfer function for each sector compensates for distortions in the unread image, and determining the meanings of the corrected locations of the marks within each sector of the unread image by reference to corresponding locations within each sectors of the model image.
3. The process of claim 2, wherein the sector is defined by the location of three alignment marks and the process further comprises the step of: applying a three point transfer function equation to each sector defined by three alignment marks.
4. The process of claim 3, further comprising the steps of: constructing an x, y coordinate system for the location of marks in the model image and an x', y' coordinate system for marks in the unread image, wherein the three mark transfer function is: x' = A'x + B'y + C and y' = D'x + E'y = F'.
5. The process of claim 2, wherein the sector is defined by the location of four alignment marks and the process further comprises the step of: applying a four point transfer function equation to each sector defined by four alignment marks.
6. The process of claim 5, further comprising the steps of: constructing an x, y coordinate system for the location of marks in the model form image and an x', y1 coordinate system for marks in the unread form image, wherein the four mark transfer function is: and
Figure imgf000021_0001
Figure imgf000021_0002
7. The process of claim 2, further comprising the steps of applying a multi-mark transfer function to each sector of the unread form image defined by eight alignment marks, wherein the transfer function is
Figure imgf000021_0003
and
Figure imgf000021_0004
8. The process of claim 1 , further comprising the steps of: generating corrected locations of specific marks on the unread image, wherein the locations of the specific marks on the model image are known; calculating a difference between the corrected locations of the specific marks on the unread image to the corresponding known locations of the specific marks on the model image; establishing an error threshold; and comparing the difference to the threshold, and rejecting the form if the difference exceeds the threshold.
9. A system for processing and correcting markings on a known form, the system comprising: lighting that illuminates a form that reflects some of the light; a camera arranged to receive the reflected light from the form; camera electronics to extract and store the unread image of the form, the unread image having alignment marks at locations; a model image of the form having alignment marks; sectors defined by the locations of the alignment marks; a transfer functions for each sector from the locations of corresponding alignment marks of the stored unread and model images; and an application that generates corrected locations of marks within each sector of the stored unread image by application of the transfer function to the marks within each sector of the stored unread image.
10. The system of claim 9, further comprising: a coordinate system for locations on the stored unread and the model images; and an application that determines the meanings of marks with corrected location by reference to the corresponding locations on the model image.
11. The system of claim 10, further comprising: specific marks with corrected locations on the stored unread image and corresponding specific marks on the model image, wherein the locations of the specific marks on the model image are known; a calculator for calculating a difference between the corrected locations of the specific marks on the unread image to the corresponding known locations of the specific marks on the model image; an error threshold; and a comparator that compares the difference to the error threshold, and rejects the form if the difference exceeds the threshold.
12. A computer readable medium containing executable program instructions for determining the location of marks on a known type of form, the executable program comprising one or more program instructions for: storing an image of a model of the unread form in a computer system, the model image containing alignment marks; storing an image of the unread form in the computer system, the unread image containing alignment marks that correspond to the alignments marks on the model image; partitioning the model image into a group of sectors, wherein each sector is defined by the locations of at least three alignment marks; detecting the sectors on the unread image that correspond to the sectors on the model image by comparing the alignment marks on the model image to the corresponding alignment marks on the unread image; generating a transfer function for each sector of the unread image from the locations of corresponding alignments marks of the model and unread images; and applying each sector's transfer function to the locations of marks on the corresponding sectors of the unread image to generate corrected locations of the marks, thereby enabling the reading of the unread form.
13. The computer readable medium of claim 12, further comprising program instructions for: forming a coordinate system for each sector, wherein the transfer function for each sector compensates for distortions, determining the meanings of the corrected locations of the marks within each sector of the unread image by reference to corresponding locations of marks within each sectors of the model image; and applying a three mark transfer function to each sector defined by three alignment marks.
14. The computer readable medium of claim 13 , wherein the three mark transfer function is: x' = A'x + B'y + C; and y1 = D'x + E'y = F';
15. The computer readable medium of claim 12, further comprising program instructions for: forming a coordinate system for each sector, wherein the transfer function for each sector compensates for distortions, determining the meanings of the corrected locations of the marks within each sector of the unread image by reference to corresponding locations of marks within each sectors of the model image; and applying a three mark transfer function to each sector defined by four alignment marks.
16. The computer readable medium of claim 15 , wherein the four mark transfer function is: and
Figure imgf000024_0001
Figure imgf000024_0002
17. The computer readable medium of claim 12, further comprising program instructions for: forming a coordinate system for each sector; wherein the transfer function for each sector compensates for distortions, determining the meanings of the corrected locations of the marks within each sector of the unread image by reference to corresponding locations of marks within each sectors of the model image; and applying a multi-mark transfer function to each sector defined by eight alignment marks, wherein the transfer function is
Figure imgf000025_0001
and
Figure imgf000025_0002
18. The computer readable medium of claim 12, further comprising program instructions for: determining the meanings of the corrected locations of the marks within each sector of the unread image by reference to corresponding locations of marks within each sectors of the model image.
19. The process of claim 1 , wherein the unread form is selected from the group consisting of a printed document, a play slip, a lottery scratch ticket and a lottery instant ticket.
20. The system of claim 9, wherein the known form is selected from the group consisting of a printed document, a play slip, a lottery scratch ticket and a lottery instant ticket.
PCT/US2009/065660 2008-11-26 2009-11-24 Image processing for curvature correction WO2010062884A1 (en)

Priority Applications (7)

Application Number Priority Date Filing Date Title
AU2009319844A AU2009319844A1 (en) 2008-11-26 2009-11-24 Image processing for curvature correction
EP09796868A EP2366171A1 (en) 2008-11-26 2009-11-24 Image processing for curvature correction
CA2744698A CA2744698A1 (en) 2008-11-26 2009-11-24 Image processing for curvature correction
GB1110263A GB2477707A (en) 2008-11-26 2009-11-24 Image Processing For Curvature Correction
JP2011538665A JP2012510235A (en) 2008-11-26 2009-11-24 Image processing for curve correction
CN2009801529378A CN102265303A (en) 2008-11-26 2009-11-24 image processing method for curvature correction
IL213149A IL213149A0 (en) 2008-11-26 2011-05-26 Image processing for curvature correction

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/323,701 US8233200B2 (en) 2005-03-14 2008-11-26 Curvature correction and image processing
US12/323,701 2008-11-26

Publications (1)

Publication Number Publication Date
WO2010062884A1 true WO2010062884A1 (en) 2010-06-03

Family

ID=41664897

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2009/065660 WO2010062884A1 (en) 2008-11-26 2009-11-24 Image processing for curvature correction

Country Status (10)

Country Link
US (1) US8233200B2 (en)
EP (1) EP2366171A1 (en)
JP (1) JP2012510235A (en)
KR (1) KR20110102361A (en)
CN (1) CN102265303A (en)
AU (1) AU2009319844A1 (en)
CA (1) CA2744698A1 (en)
GB (1) GB2477707A (en)
IL (1) IL213149A0 (en)
WO (1) WO2010062884A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2713315A1 (en) 2012-09-26 2014-04-02 Lotsys Method and device for reading a document

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100160022A1 (en) * 2008-12-24 2010-06-24 Gtech Corporation Flexible self-describing wagering game entries
US8396301B2 (en) 2009-09-24 2013-03-12 Gtech Corporation System and method for document location and recognition
TWI516119B (en) * 2011-01-25 2016-01-01 華晶科技股份有限公司 Electronic apparatus, image capturing apparatus and method thereof
JP2015126457A (en) * 2013-12-27 2015-07-06 ブラザー工業株式会社 Program for server device, server device, and remote conference method
JP2015126456A (en) 2013-12-27 2015-07-06 ブラザー工業株式会社 Program for communication terminal device, communication terminal device, program for server device, and remote conference system
CN104921698B (en) * 2015-07-13 2016-08-24 中北大学 A kind of method revising corneal curvature instrument system astigmatism
CN110027342A (en) * 2017-12-20 2019-07-19 卡西欧计算机株式会社 Paper is shot and device, method for correcting image, leaf turner and the image pickup method of electronization
US10746667B2 (en) * 2018-11-27 2020-08-18 General Electric Company Fluorescent penetrant inspection system and method
KR102279259B1 (en) * 2020-01-17 2021-07-20 네이버 주식회사 Method for compensating image distortion and, method for collecting objects from the image by using the same
US11948479B2 (en) * 2020-03-18 2024-04-02 Nec Corporation Of America Tampering detection based on non-reproducible marks in a tampering evident element
CN112581523A (en) * 2020-12-01 2021-03-30 广东世宇科技股份有限公司 Method and device for determining ball cake position, electronic equipment and readable storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6954290B1 (en) * 2000-11-09 2005-10-11 International Business Machines Corporation Method and apparatus to correct distortion of document copies

Family Cites Families (90)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4141036A (en) 1977-03-10 1979-02-20 General Electric Company Solid state color camera
US4240748A (en) 1978-06-26 1980-12-23 Caere Corporation Hand-held optical character recognition wand with visual aligner
US5668364A (en) 1985-02-28 1997-09-16 Symbol Technologies, Inc. Target finder in electro-optical scanners
US4760247A (en) 1986-04-04 1988-07-26 Bally Manufacturing Company Optical card reader utilizing area image processing
US4724307A (en) 1986-04-29 1988-02-09 Gtech Corporation Marked card reader
US4736109A (en) 1986-08-13 1988-04-05 Bally Manufacturing Company Coded document and document reading system
US4933778A (en) * 1988-09-19 1990-06-12 Eastman Kodak Company Calibration of platen registration edges in copiers
US5039847A (en) 1989-03-24 1991-08-13 Omron Corporation Image processing apparatus having an image scanner
US5258605A (en) 1990-03-13 1993-11-02 Symbol Technologies, Inc. Scan generators for bar code reader using linear array of lasers
US5216595A (en) 1990-03-20 1993-06-01 Ncr Corporation System and method for integration of lottery terminals into point of sale systems
US5103490A (en) * 1990-06-13 1992-04-07 National Computer Systems, Inc. Method and apparatus for storing and merging multiple optically scanned images
US5239165A (en) 1991-04-11 1993-08-24 Spectra-Physics Scanning Systems, Inc. Bar code lottery ticket handling system
US5262624A (en) 1991-07-31 1993-11-16 National Computer Systems, Inc. Opposite surface scanning of a mark sense form
US5452379A (en) 1991-09-13 1995-09-19 Meadowbrook Industries, Ltd. Image capture and storage techniques in association with optical mark reading
US5214470A (en) * 1992-04-09 1993-05-25 Xerox Corporation Method and apparatus for compensating for dirt or etched areas on a document platen
US6150930A (en) 1992-08-14 2000-11-21 Texas Instruments Incorporated Video equipment and method to assist motor vehicle operators
US7168084B1 (en) * 1992-12-09 2007-01-23 Sedna Patent Services, Llc Method and apparatus for targeting virtual objects
US6267670B1 (en) 1997-03-21 2001-07-31 Walker Digital, Llc System and method for performing lottery ticket transactions utilizing point-of-sale terminals
US5417424A (en) 1993-09-28 1995-05-23 Gtech Corporation Player operated win checker appended to lottery agent terminal
US5735094A (en) * 1994-01-03 1998-04-07 Ultra-Tex Surfaces, Inc. Method for producing an ornamental concrete surface
DE69521153T2 (en) * 1994-01-20 2002-05-02 Omron Tateisi Electronics Co Image processing device and method for detecting a reference pattern
US7387253B1 (en) 1996-09-03 2008-06-17 Hand Held Products, Inc. Optical reader system comprising local host processor and optical reader
US5362105A (en) 1994-03-18 1994-11-08 Scott Arthur C Scanner
US6053405A (en) 1995-06-07 2000-04-25 Panda Eng., Inc. Electronic verification machine for documents
US6875105B1 (en) 1994-06-22 2005-04-05 Scientific Games Inc. Lottery ticket validation system
US5834749A (en) 1994-08-30 1998-11-10 Durbin; Dennis A. Optical image capture system for reading targets at oblique angles
US5945655A (en) 1994-09-22 1999-08-31 Gilgeous; Earle Apparatus and method for counting bingo cards
ES2116779T3 (en) 1994-10-25 1998-07-16 United Parcel Service Inc PROCEDURE AND PORTABLE DEVICE FOR CONTACTLESS READING OF AN IMAGE ON A LABEL.
US5673125A (en) 1995-08-03 1997-09-30 Agfa Division, Bayer Corporation Scanning system for scanning transmissive and reflective originals
US5818026A (en) 1996-02-06 1998-10-06 Symbol Technologies, Inc. Hand-marked lottery selection slip scanning system and method
US6366696B1 (en) 1996-12-20 2002-04-02 Ncr Corporation Visual bar code recognition method
US20010029513A1 (en) 1997-03-05 2001-10-11 Hideyuki Kuwano Integrated apparatus and system for storing, retrieving, and transmitting documents using document IDs and document ID marks
US6223986B1 (en) 1997-04-17 2001-05-01 Psc Scanning, Inc. Aiming aid for optical data reading
US20040209665A1 (en) 1997-08-15 2004-10-21 Walker Jay S. Products and processes for applying conditions to a lottery entry
US7028899B2 (en) 1999-06-07 2006-04-18 Metrologic Instruments, Inc. Method of speckle-noise pattern reduction and apparatus therefore based on reducing the temporal-coherence of the planar laser illumination beam before it illuminates the target object by applying temporal phase modulation techniques during the transmission of the plib towards the target
JP3461272B2 (en) 1997-09-22 2003-10-27 キヤノン株式会社 Image reading method and apparatus
US6561428B2 (en) 1997-10-17 2003-05-13 Hand Held Products, Inc. Imaging device having indicia-controlled image parsing mode
US6086477A (en) 1998-03-31 2000-07-11 Walker Digital, Llc Methods and apparatus wherein a lottery entry is entered into lottery drawings until the lottery entry is identified as a winner
US6262670B1 (en) 1998-04-10 2001-07-17 Kevin R. Ballou Source measure unit having secondary feedback for eliminating transients during range changing
US6199044B1 (en) 1998-05-27 2001-03-06 Intermec Ip Corp. Universal data input and processing device, such as universal point-of-sale device for inputting and processing bar code symbols, document images, and other data
US6157435A (en) 1998-05-29 2000-12-05 Eastman Kodak Company Image processing
US6340114B1 (en) 1998-06-12 2002-01-22 Symbol Technologies, Inc. Imaging engine and method for code readers
US6176429B1 (en) 1998-07-17 2001-01-23 Psc Scanning, Inc. Optical reader with selectable processing characteristics for reading data in multiple formats
US6357658B1 (en) 1999-04-28 2002-03-19 Peripheral Dynamics, Inc. Apparatus and methods for scanning documents including OMR, bar-code, and image data
US6470099B1 (en) * 1999-06-30 2002-10-22 Hewlett-Packard Company Scanner with multiple reference marks
EP1067362A1 (en) 1999-07-09 2001-01-10 Hewlett-Packard Company Document imaging system
US6234899B1 (en) 1999-07-20 2001-05-22 Scientific Games, Inc. System for linking a unique identifier to an instant game ticket
US20020010022A1 (en) 1999-08-25 2002-01-24 Mindy Vincent Intuitive system and method for entering a wager transaction
US6760476B2 (en) 2000-02-07 2004-07-06 Martin G. Meder Method of checking lottery ticket numbers
US6494372B2 (en) 2000-04-04 2002-12-17 International Business Machines Corporation Self service terminal and method for processing transaction forms
US6816630B1 (en) * 2000-05-22 2004-11-09 Electro-Sensors, Inc. System and method for creating and processing data forms
DE60041580D1 (en) 2000-07-11 2009-04-02 Datalogic Spa Device and optical element for viewing and visual display of a read-out area of a code reader
US20020047047A1 (en) 2000-09-06 2002-04-25 Paul Poloniewicz Zero-footprint camera-based point-of-sale bar code presentation scanning system
US20020048403A1 (en) 2000-10-24 2002-04-25 Electronic Warfare Associates, Inc Mark recognition system and method for identification of one or more marks on an object
US7490774B2 (en) 2003-11-13 2009-02-17 Metrologic Instruments, Inc. Hand-supportable imaging based bar code symbol reader employing automatic light exposure measurement and illumination control subsystem integrated therein
TW512621B (en) 2001-02-15 2002-12-01 Benq Corp Photography negative scanning device using cold cathode tube to unify infrared
US7621810B2 (en) 2001-02-27 2009-11-24 Scientific Games International, Inc. System and method for selling lottery game tickets through a point of sale system
US6650427B2 (en) 2001-06-01 2003-11-18 Transact Technologies Incorporated Self validating printer with ticket voiding features
US6722569B2 (en) 2001-07-13 2004-04-20 Welch Allyn Data Collection, Inc. Optical reader having a color imager
KR100460336B1 (en) 2001-07-26 2004-12-04 김택진 Reading unit for optical mark reader
GB2378073B (en) * 2001-07-27 2005-08-31 Hewlett Packard Co Paper-to-computer interfaces
US6736319B2 (en) 2001-09-04 2004-05-18 Renaissance Learning, Inc. Optical mark reader
US7077313B2 (en) 2001-10-01 2006-07-18 Avante International Technology, Inc. Electronic voting method for optically scanned ballot
US7070115B2 (en) * 2002-01-23 2006-07-04 Debiase Salvatore E Ballot form and method for making and using same
US20030186734A1 (en) 2002-03-12 2003-10-02 Lemay Steven G. Gaming machine including a lottery ticket dispenser
US7024375B2 (en) 2002-06-17 2006-04-04 Chau Wayne P System and method for purchasing game and lottery tickets
US7060968B1 (en) 2002-06-04 2006-06-13 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Method and apparatus for optical encoding with compressible imaging
JP2004172854A (en) 2002-11-19 2004-06-17 Seiko Epson Corp Image sensor controller, electronic apparatus and image sensor control method
DE10303166A1 (en) 2003-01-27 2004-08-05 Wincor Nixdorf International Gmbh lottery terminal
DE10315442A1 (en) 2003-04-03 2004-11-11 Bts Media Solutions Gmbh Process and circuit for scaling raster images
KR100429703B1 (en) 2003-04-07 2004-05-03 주식회사 아이콘랩 Method for capturing of real-time image
US6929184B2 (en) 2003-11-05 2005-08-16 Symbol Technologies, Inc. Monitoring bi-directional motor drive failure in electro-optical reader
US7298902B2 (en) * 2004-01-20 2007-11-20 Educational Testing Service Method and system for performing image mark recognition
US7532755B2 (en) 2004-01-22 2009-05-12 Lexmark International, Inc. Image classification using concentration ratio
US7090131B2 (en) 2004-02-16 2006-08-15 Seiko Epson Corporation Single-pass magnetic reading and optical reading apparatus and method
US20050237580A1 (en) 2004-04-16 2005-10-27 Dave Coleman Scanner read head for images and optical mark recognition
US20050238260A1 (en) 2004-04-16 2005-10-27 Dave Coleman Image and optical mark scanner with encryption
US20060063589A1 (en) 2004-09-17 2006-03-23 Nokia Corporation Method and system for providing a personal message relating to a personalized item that carries on a surface a personalized optical code
US7247095B2 (en) 2004-10-08 2007-07-24 Nulph James L Method and system for marketing and game selection for lottery products
US7206008B2 (en) * 2004-10-28 2007-04-17 Xerox Corporation Method for calibrating color in a printing device
US20060221354A1 (en) 2005-04-01 2006-10-05 Charles Slaten Automatic printing device selection based on human-readable form submission
US20060290886A1 (en) 2005-05-24 2006-12-28 Mr. Dario Santos Digital Capturing Pharmaceutical System
US7593144B2 (en) 2005-06-28 2009-09-22 Xerox Corporation Controlling scanning and copying devices through implicit gestures
US7755808B2 (en) 2005-11-17 2010-07-13 Xerox Corporation Document scanner dust detection systems and methods
US20090167960A1 (en) 2006-03-29 2009-07-02 Pioneer Corporation Picture processing apparatus
US8064728B2 (en) 2007-03-30 2011-11-22 Intel Corporation Traceability marks
KR101279034B1 (en) 2007-07-11 2013-07-02 삼성전자주식회사 Scanner module and image scanning apparatus
US7843611B2 (en) 2007-07-18 2010-11-30 Kuwait University High speed flatbed scanner comprising digital image-capture module with two-dimensional optical image photo-sensor or digital camera
JP4428424B2 (en) 2007-08-20 2010-03-10 ソニー株式会社 Information processing apparatus, information processing method, program, and recording medium
US8014035B2 (en) 2008-09-10 2011-09-06 Xerox Corporation Decoding message data embedded in an image print via halftone dot orientation

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6954290B1 (en) * 2000-11-09 2005-10-11 International Business Machines Corporation Method and apparatus to correct distortion of document copies

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
BOONE J M ET AL: "Analysis and correction of imperfections in the image intensifier-TV-digitizer imaging chain", MEDICAL PHYSICS, AIP, vol. 18, no. 2, 1 March 1991 (1991-03-01), MELVILLE, NY, US, pages 236 - 242, XP000228810, ISSN: 0094-2405 *
FANTOZZI SILVIA ET AL: "A global method based on thin-plate splines for correction of geometric distortion: An application to fluoroscopic images", MEDICAL PHYSICS, AIP, vol. 30, no. 2, 1 February 2003 (2003-02-01), MELVILLE, NY, US, pages 124 - 131, XP012011975, ISSN: 0094-2405 *
GLASBEY C. A. ET AL: "A review of image-warping methods", JOURNAL OF APPLIED STATISTICS, vol. 25, no. 2, 1 April 1998 (1998-04-01), pages 1 - 18, XP002570074, DOI: 10.1080/02664769823151 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2713315A1 (en) 2012-09-26 2014-04-02 Lotsys Method and device for reading a document

Also Published As

Publication number Publication date
GB201110263D0 (en) 2011-08-03
US20090074327A1 (en) 2009-03-19
US8233200B2 (en) 2012-07-31
AU2009319844A1 (en) 2010-06-03
GB2477707A (en) 2011-08-10
CN102265303A (en) 2011-11-30
JP2012510235A (en) 2012-04-26
EP2366171A1 (en) 2011-09-21
CA2744698A1 (en) 2010-06-03
IL213149A0 (en) 2011-07-31
KR20110102361A (en) 2011-09-16

Similar Documents

Publication Publication Date Title
US8233200B2 (en) Curvature correction and image processing
US20110019243A1 (en) Stereoscopic form reader
EP1235181A2 (en) Improvements relating to document capture
US10713528B2 (en) System for determining alignment of a user-marked document and method thereof
US8322620B2 (en) Decoding distorted symbols
EP0701225A2 (en) System for transcribing images on a board using a camera based board scanner
US7991244B2 (en) Variable skew correction system and method
US8055901B2 (en) Optical signature to enable image correction
WO2011129113A1 (en) Form processing device, form processing method, and recording medium
CN109558844A (en) The method and apparatus of self-defined template discrimination is promoted based on image normalization
CN111860527A (en) Image correction method, image correction device, computer device, and storage medium
US8913852B2 (en) Band-based patch selection with a dynamic grid
KR20090062115A (en) Apparatus and method for media image detection, and system with the same
CN111862082A (en) Train brake pad thickness rechecking method and system thereof
JP7262927B2 (en) Image processing device, image processing method and program
KR101766787B1 (en) Image correction method using deep-learning analysis bassed on gpu-unit
CN112997217A (en) Document detection from video images
JP2009025992A (en) Two-dimensional code
WO2009097125A1 (en) Recognition of scanned optical marks for scoring student assessment forms
JP7197009B2 (en) Object management system
WO2022244616A1 (en) Inspection device, computer program, and inspection method
JP3258828B2 (en) Optical disc identification symbol recognition method
CN115830134A (en) Camera calibration method and device, electronic equipment and medium
WO1996022578A1 (en) Method and apparatus for characterizing the performance of a printer or an imaging system
AU2011203230A1 (en) Variable patch size alignment hints

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200980152937.8

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09796868

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2744698

Country of ref document: CA

Ref document number: 2011538665

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 213149

Country of ref document: IL

WWE Wipo information: entry into national phase

Ref document number: 2009319844

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 4226/CHENP/2011

Country of ref document: IN

ENP Entry into the national phase

Ref document number: 2009319844

Country of ref document: AU

Date of ref document: 20091124

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 1110263

Country of ref document: GB

Kind code of ref document: A

Free format text: PCT FILING DATE = 20091124

WWE Wipo information: entry into national phase

Ref document number: 1110263.9

Country of ref document: GB

Ref document number: 2009796868

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 20117014521

Country of ref document: KR

Kind code of ref document: A