WO2005045655A2 - Method for determining location on display surface and interactive display system - Google Patents

Method for determining location on display surface and interactive display system Download PDF

Info

Publication number
WO2005045655A2
WO2005045655A2 PCT/JP2004/016625 JP2004016625W WO2005045655A2 WO 2005045655 A2 WO2005045655 A2 WO 2005045655A2 JP 2004016625 W JP2004016625 W JP 2004016625W WO 2005045655 A2 WO2005045655 A2 WO 2005045655A2
Authority
WO
WIPO (PCT)
Prior art keywords
patterns
sequence
display surface
light
intensities
Prior art date
Application number
PCT/JP2004/016625
Other languages
French (fr)
Other versions
WO2005045655A3 (en
Inventor
Paul H. Dietz
Darren L. Leigh
Ramesh Raskar
Johnny Chung Lee
Original Assignee
Mitsubishi Denki Kabushiki Kaisha
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mitsubishi Denki Kabushiki Kaisha filed Critical Mitsubishi Denki Kabushiki Kaisha
Priority to JP2006519005A priority Critical patent/JP4675889B2/en
Priority to DE602004031140T priority patent/DE602004031140D1/en
Priority to EP04799563A priority patent/EP1680732B1/en
Publication of WO2005045655A2 publication Critical patent/WO2005045655A2/en
Publication of WO2005045655A3 publication Critical patent/WO2005045655A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • G06F3/0317Detection arrangements using opto-electronic means in co-operation with a patterned surface, e.g. absolute position or relative movement detection for an optical mouse or pen positioned with respect to a coded surface
    • G06F3/0321Detection arrangements using opto-electronic means in co-operation with a patterned surface, e.g. absolute position or relative movement detection for an optical mouse or pen positioned with respect to a coded surface by optically sensing the absolute position with respect to a regularly patterned surface forming a passive digitiser, e.g. pen optically detecting position indicative tags printed on a paper sheet
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03542Light pens for emitting or receiving light

Definitions

  • This invention relates generallyto input devices for interaction with display systems, and more particularly to light pens used with such displays.
  • Light pens are often used as input devices with graphic user interfaces, see Myers, "A Brief History of Human Computer Interaction Technology," ACM Interactions, Vol. 5, no. 2, pp. 44-54, March 1998, for a brief history. Although light pens have been used since 1954 , it was the work of Ivan Sutherland at Lincoln Laboratories, that popularized the use of light pens with computers, Sutherland, "Sketchpad: The First Interactive Computer Graphics," Ph.D. Thesis, MIT, 1963.
  • a light pen As an input device, a light pen has some advantages over a mouse.
  • the mouse is an indirect input device. Movement of the mouse on a horizontal surface causes movement of a cursor on a vertical display. The amount of movement, and speed of movement of the mouse and cursor can be different. In contrast, the pointing of a light pen is direct, and thus more intuitive.
  • the typical light pen is used with a display system as follows. Most CRT-based displays are raster scanned. Individualpixels onthe display are illuminatedatpredetermined times depending on the location of the pixel in the scanning order. By detecting when a pixel is illuminated, the location of the pixel on the display surface can be determined. Therefore, the typical light pen includes a photo-sensor connected to a timing circuit. The timing of the sensed pixel, synchronized to the raster scanning, reveals its location. Typically, the light pen is held very close, if not in contact with screen.
  • the invention projects imperceptible patterns of structured light on a pixel-based display device.
  • the patterns can be black and white, gray scale, or full red, green, blue (RGB) colors. Colored patterns can convey more information than binary black andwhitepatterns .
  • Thepatterns are interleavedwithperceptible content .
  • the patterns are arranged so as to encode location information. While these patterns are imperceptible by the human visual system, the patterns can be detected by photo-sensors. Because the encodings are unique for all pixel locations, it is possible to decode the patterns to determine location information.
  • the patterns of structured light can be spatial, temporal, or spatio-temporally unique at each location.
  • the encoding is completely in the time-domain.
  • the pattern of light intensity at each pixel, over time encodes the location information.
  • the location is encoded by unique intensity variations over a neighboring region of pixels. Therefore, a neighborhood of pixels is sampled in order to determine a pixel location.
  • This embodiment uses multiple photo-sensors.
  • the spatial implementation can determine the location from a single pattern.
  • the patterns vary over space and time. This technique can use a smaller neighborhood then in the spatial case and fewer patterns than in the temporal case . Thus, cost and speed can be traded-off to optimize the system for a particular application.
  • the light can be at infrared (IR) or other frequencies out of the range of human vision.
  • IR infrared
  • an infrared filter can be added to the color wheel used by such devices.
  • Asecondtechnique uses balancedmasking .
  • the light is in the visible range.
  • a total sum of light to all pixels is identical over time.
  • each pattern is immediately followed by a reverse or negative of the pattern .
  • the twopatterns form a continuous gray tone. If the patterns are displayed for a relatively small amount of time, the onlyperceptible artifact is a slight decrease in the contrast ratio.
  • the structured light patterns are detected by a photo-sensor.
  • the detected patterns are decoded to determine the location information.
  • a light pen with a single photo-sensor yields sufficient information for location estimation.
  • an array of photo-sensors e.g., a camera, is used to detect light intensity variations for the neighborhood of pixels.
  • Figure 1 is a block diagram of a display and light pen system according to the invention.
  • Figure 2 is a flow diagram of a process for generating imperceptible patterns according to the invention.
  • Figure 3 is a flow diagram of a process for decoding patterns according to the invention to obtain location information
  • Figure 4 are patterns according to the invention.
  • Figure 5 is a block diagram of a display mosaic
  • Figure 6 is a diagram of a display system with a curved surface.
  • Figure 1 shows a light pen system 100 for a pixel-based display.
  • the system includes a pattern and content generator 110 coupled to an image generator 120, e.g., a projector.
  • the image generator renders a sequence of patterns 111 on a display surface 130.
  • the display surface is pixel-based, instead of raster scanned, although the system also works with raster-scanned displays.
  • the system also includes a light pen
  • the task it to determine 2D coordinates of an arbitrary location 101 on the display surface 130.
  • the image generator 120 can use a LCD screen, a digital-mirror array, liquid crystal on silicon (LCOS) , and organic LED technologies in either front or rear proj ection modes . It should be noted that the invention can also be used with conventional CRT-based displays.
  • the light pen can use a single photo-sensor
  • the light pen can include a pressure sensitive switch, which is activated when the sensor is pressed against the display surface. This prevents spurious readings .
  • the patterns 111 are projected using infrared (IR) light. This can be accomplished by passing light through a condensing lens, an infrared color filter mounted on a color wheel, a shaping lens, to a digital mirror device (DMD) , and then through a proj ector lens onto the display surface 130.
  • IR infrared
  • DMD digital mirror device
  • the sequence of patterns 111 is generated by changing sequentially states of the mirrors. Because the patterns are detected by the photo-sensor 141, rather than the human visual system, the patterns can be generated at an extremely high rate, e.g., greater than 1 KHz. It is known that the human visual system is sensitive to stimuli only in a certain temporal window, called the window of visibility, see Watson et al., "Window of Visibility: apsychophysical theory of fidelity in time-sampled visual motion displays," J. Opt. Soc. Am. A, Vol.3, No. 3, pp. 300-307, March 1986. The human visual system cannot comprehend images beyond a certain temporal frequency limit. The invention generates on the order of a hundred unique patterns in one millisecond, which is well outside the temporal window of visibility.
  • the sequence of patterns is generated 200 repeatedly 240 and has the following distinct parts.
  • a headerpattern is generated 210.
  • the header pattern is used to indicate the start of the sequence, the rate at which the patterns are generated.
  • the header can also be used to calibrate the light pen to the relative light and dark intensities of the patterns .
  • the header can also identify the projector.
  • Gray code 220 andvertical Gray code 230 patterns are generated, see Gray, "Pulse Code Communication", U. S. Patent 2,632,058, March 17, 1953. These patterns have a unique temporal sequence of light intensities for each individual pixel of the display 130. If groups of pixels are sensed, then each group has a unique pattern. That way, sensing the light intensity at a particular pixel over time, and decoding the pattern can determine the 2D coordinates of the location 101 of the pixel.
  • Gray codes are not the only way to achieve this property, Gray codes have other benefits. Most important, Gray code patterns guarantee that edges in succeeding patterns do not line up. This minimizes ambiguity in the case where the light pen 140 also senses parts of neighboring pixels.
  • Figure 4 shows a sequence of patterns 400 for a 4x4 pixel array.
  • the first pattern 401 is all light
  • the second pattern 402 is the reverse of the first pattern, i.e., all dark.
  • This pair of patterns is the header and enables the light pen to synchronize to the start of the sequence.
  • This header pair of patterns can also give reference levels of light intensity for calibration purposes by averaging the patterns.
  • the length of time the header patterns are displayed indicates the timing of the patterns.
  • the pattern generator 110 and the decoder 150 do not need to be synchronized with each other.
  • the header patterns can be repeated in an arbitrary order to generate a binary signal fromthe all light anddarkpatterns, e.g.01011101, where the first four ⁇ bits' are a start of sequence indicator, and the next n are other information.
  • Eachfollowingpair 403-404, 405-406, 407-408 divides the display inhalf, downto thepixel level, first for the horizontal version, and then for the vertical version so that the adjacency property of the Gray code is met.
  • the light pen is inactive most of the time while the underlying perceptible content is displayed.
  • Figure 3 shows the location decodingmethod 300. First, the header is detected 310. Then, the horizontal intensity values are measured 320, followed by measuring 330 the vertical intensity values. From these measured values, the coordinates of the location 101 are determined 340.
  • each pattern image can be extremely short, on the order of 10 microseconds.
  • the sequence of patterns can be very short, e.g., less than 1 millisecond.
  • adding the sequence 111 has minimal impact on an overall brightness of the projected display. In fact, the impact is so small, that the sequences can be displayed at a higher rate to increase the update rate of the location information.
  • the location information is encoded in red, green or blue (RGB) patterns. This is done in a way so that the patterns remain imperceptible.
  • RGB red, green or blue
  • One way of decreasing the number of patterns in the sequence is to enlarge the sensing area by using multiple sensors in the light pen 140.
  • the sequence 111 has a single pattern.
  • the light pen has as at least as many sensors as the number of patterns in the temporal sequence, i.e., twenty-two .
  • two projectors are used, a first projectordisplays the imperceptible patterns 111, while a second projector 121 displays the underlying perceptible content 122.
  • a third projector can interleave a second sequence of patterns so that 3D coordinates of the location information can be obtained in the case where the display surface 610 is curved as shown in Figure 6, and as described in U.S. Patent Application Sn. 10/394,315, "Method and System for Displaying Images on Curved Surfaces," filed by Raskar et al . , on March 21, 2003, incorporated herein by reference.
  • the display is a mosaic of partially overlapping images 501 to generate a largerpanoramic images 510 as described inU. S .
  • each header sequence 210 can include an identification so that the decoder can differentiate the different sequences of patterns projected by multiple projectors .
  • the invention has advantages over resistive touch screens, which are relatively expensive and can only distinguish a single touched location, at anyone time.
  • Conventional vision-based systems are also more complex to implement and those type of systems have a problem with shadowing and accuracy.

Abstract

An interactive display system includes a pixel-based display surface and a light pen. A sequence of patterns is projected onto the display surface. The sequence of patterns has a unique sequence of light intensities for each location of the display surface. Intensities of light at an arbitrary location are sensed by the light pen while projecting the sequence of patterns. The intensities of light are decoded to determine coordinates of the arbitrary location.

Description

DESCRIPTION
Method for Determining Location on Display Surface and Interactive Display System
Technical Field
This inventionrelates generallyto input devices for interaction with display systems, and more particularly to light pens used with such displays.
Background Art
Light pens are often used as input devices with graphic user interfaces, see Myers, "A Brief History of Human Computer Interaction Technology," ACM Interactions, Vol. 5, no. 2, pp. 44-54, March 1998, for a brief history. Although light pens have been used since 1954 , it was the work of Ivan Sutherland at Lincoln Laboratories, that popularized the use of light pens with computers, Sutherland, "Sketchpad: The First Interactive Computer Graphics," Ph.D. Thesis, MIT, 1963.
As an input device, a light pen has some advantages over a mouse. The mouse is an indirect input device. Movement of the mouse on a horizontal surface causes movement of a cursor on a vertical display. The amount of movement, and speed of movement of the mouse and cursor can be different. In contrast, the pointing of a light pen is direct, and thus more intuitive. In the prior art, the typical light pen is used with a display system as follows. Most CRT-based displays are raster scanned. Individualpixels onthe display are illuminatedatpredetermined times depending on the location of the pixel in the scanning order. By detecting when a pixel is illuminated, the location of the pixel on the display surface can be determined. Therefore, the typical light pen includes a photo-sensor connected to a timing circuit. The timing of the sensed pixel, synchronized to the raster scanning, reveals its location. Typically, the light pen is held very close, if not in contact with screen.
Unfortunately, the traditional scanned CRT-based displays are being replaced by pixel-oriented displays, such as LCD screens, digital-mirror arrays, and organic LED technologies. These pixel-based displays are not scanned, and thus the traditional light pen technique based on the timing of the raster scan is not applicable for these newer display modalities.
Therefore, there is a need for a system andmethod that interfaces a light pen with a pixel-based display.
Disclosure of Invention
The invention projects imperceptible patterns of structured light on a pixel-based display device. The patterns can be black and white, gray scale, or full red, green, blue (RGB) colors. Colored patterns can convey more information than binary black andwhitepatterns . Thepatterns are interleavedwithperceptible content .
The patterns are arranged so as to encode location information. While these patterns are imperceptible by the human visual system, the patterns can be detected by photo-sensors. Because the encodings are unique for all pixel locations, it is possible to decode the patterns to determine location information. The patterns of structured light can be spatial, temporal, or spatio-temporally unique at each location.
In the temporal embodiment, the encoding is completely in the time-domain. The pattern of light intensity at each pixel, over time, encodes the location information. Thus, it is sufficient to sample the light intensities at a single pixel over time in order to determine the location of the pixel.
In the spatial embodiment, the location is encoded by unique intensity variations over a neighboring region of pixels. Therefore, a neighborhood of pixels is sampled in order to determine a pixel location. This embodiment uses multiple photo-sensors. As an advantage over the temporal embodiment, the spatial implementation can determine the location from a single pattern. In the spatio-temporal embodiment, the patterns vary over space and time. This technique can use a smaller neighborhood then in the spatial case and fewer patterns than in the temporal case . Thus, cost and speed can be traded-off to optimize the system for a particular application.
A number of different techniques can be used to make the patterns imperceptible to the human visual system. First, the light can be at infrared (IR) or other frequencies out of the range of human vision. In digital mirror-based projectors, an infrared filter can be added to the color wheel used by such devices.
Asecondtechniqueuses balancedmasking . In this implementation, the light is in the visible range. However, a total sum of light to all pixels is identical over time. For example, each pattern is immediately followed by a reverse or negative of the pattern . When shown in quick succession, the twopatterns form a continuous gray tone. If the patterns are displayed for a relatively small amount of time, the onlyperceptible artifact is a slight decrease in the contrast ratio.
The structured light patterns are detected by a photo-sensor. The detected patterns are decoded to determine the location information.
For the temporal patterns with IR, a light pen with a single photo-sensor yields sufficient information for location estimation. For spatial patterns, an array of photo-sensors, e.g., a camera, is used to detect light intensity variations for the neighborhood of pixels.
Brief Description of Drawings
Figure 1 is a block diagram of a display and light pen system according to the invention;
Figure 2 is a flow diagram of a process for generating imperceptible patterns according to the invention;
Figure 3 is a flow diagram of a process for decoding patterns according to the invention to obtain location information;
Figure 4 are patterns according to the invention;
Figure 5 is a block diagram of a display mosaic; and
Figure 6 is a diagram of a display system with a curved surface.
Best Mode for Carrying Out the Invention
Figure 1 shows a light pen system 100 for a pixel-based display. The system includes a pattern and content generator 110 coupled to an image generator 120, e.g., a projector. The image generator renders a sequence of patterns 111 on a display surface 130. In the preferred embodiment, the display surface is pixel-based, instead of raster scanned, although the system also works with raster-scanned displays. The system also includes a light pen
140 coupled to a location decoder 150. The task it to determine 2D coordinates of an arbitrary location 101 on the display surface 130.
The image generator 120 can use a LCD screen, a digital-mirror array, liquid crystal on silicon (LCOS) , and organic LED technologies in either front or rear proj ection modes . It should be noted that the invention can also be used with conventional CRT-based displays. The light pen can use a single photo-sensor
141 or an array of, e.g., CCD, sensors as in a camera. The light pen can include a pressure sensitive switch, which is activated when the sensor is pressed against the display surface. This prevents spurious readings .
In one embodiment, the patterns 111 are projected using infrared (IR) light. This can be accomplished by passing light through a condensing lens, an infrared color filter mounted on a color wheel, a shaping lens, to a digital mirror device (DMD) , and then through a proj ector lens onto the display surface 130. These technologies are well known.
The sequence of patterns 111 is generated by changing sequentially states of the mirrors. Because the patterns are detected by the photo-sensor 141, rather than the human visual system, the patterns can be generated at an extremely high rate, e.g., greater than 1 KHz. It is known that the human visual system is sensitive to stimuli only in a certain temporal window, called the window of visibility, see Watson et al., "Window of Visibility: apsychophysical theory of fidelity in time-sampled visual motion displays," J. Opt. Soc. Am. A, Vol.3, No. 3, pp. 300-307, March 1986. The human visual system cannot comprehend images beyond a certain temporal frequency limit. The invention generates on the order of a hundred unique patterns in one millisecond, which is well outside the temporal window of visibility.
As shown in Figure 2, the sequence of patterns is generated 200 repeatedly 240 and has the following distinct parts. First, a headerpattern is generated 210. In the headerpattern, all pixels switched on and off in unison as described below. The header pattern is used to indicate the start of the sequence, the rate at which the patterns are generated. The header can also be used to calibrate the light pen to the relative light and dark intensities of the patterns . In the case of a multiple proj ector system, as described below, the header can also identify the projector.
Next, horizontal Gray code 220 andvertical Gray code 230 patterns are generated, see Gray, "Pulse Code Communication", U. S. Patent 2,632,058, March 17, 1953. These patterns have a unique temporal sequence of light intensities for each individual pixel of the display 130. If groups of pixels are sensed, then each group has a unique pattern. That way, sensing the light intensity at a particular pixel over time, and decoding the pattern can determine the 2D coordinates of the location 101 of the pixel.
While Gray codes are not the only way to achieve this property, Gray codes have other benefits. Most important, Gray code patterns guarantee that edges in succeeding patterns do not line up. This minimizes ambiguity in the case where the light pen 140 also senses parts of neighboring pixels.
Figure 4 shows a sequence of patterns 400 for a 4x4 pixel array. The first pattern 401 is all light, the second pattern 402 is the reverse of the first pattern, i.e., all dark. This pair of patterns is the header and enables the light pen to synchronize to the start of the sequence. This header pair of patterns can also give reference levels of light intensity for calibration purposes by averaging the patterns. In addition, the length of time the header patterns are displayed indicates the timing of the patterns. Thus, the pattern generator 110 and the decoder 150 do not need to be synchronized with each other. The header patterns can be repeated in an arbitrary order to generate a binary signal fromthe all light anddarkpatterns, e.g.01011101, where the first four λbits' are a start of sequence indicator, and the next n are other information.
Eachfollowingpair 403-404, 405-406, 407-408 divides the display inhalf, downto thepixel level, first for the horizontal version, and then for the vertical version so that the adjacency property of the Gray code is met. The light pen is inactive most of the time while the underlying perceptible content is displayed.
Figure 3 shows the location decodingmethod 300. First, the header is detected 310. Then, the horizontal intensity values are measured 320, followed by measuring 330 the vertical intensity values. From these measured values, the coordinates of the location 101 are determined 340.
It should be noted that the duration of each pattern image can be extremely short, on the order of 10 microseconds. For a conventional XGAresolutionprojector, only twenty-two patterns are required for the header and all location information. Thus, the sequence of patterns can be very short, e.g., less than 1 millisecond. Thus, adding the sequence 111 has minimal impact on an overall brightness of the projected display. In fact, the impact is so small, that the sequences can be displayed at a higher rate to increase the update rate of the location information.
Other embodiments are possible. For organic LED displays, adding IR emitters may be impractical. In this case, the location information is encoded in red, green or blue (RGB) patterns. This is done in a way so that the patterns remain imperceptible. One way displays each pattern for a very short time, as described above .
When large Mark' regions are displayed, it is possible that some varying light intensities are perceived. This can be remedied by a balance masking technique. In this technique, each image is immediately followed by its reverse or negative. This gives every pixel a 50% duty cycle on average while the sequence of patterns is displayed. The net perceived effect is an almost imperceptible smooth grey image, which is the average of all images in the sequence, which only results in a slight loss of contrast.
For LCD-based displays, it may be difficult to achieve the speed required for a strictly temporal solution. One way of decreasing the number of patterns in the sequence is to enlarge the sensing area by using multiple sensors in the light pen 140.
In an extreme case, the sequence 111 has a single pattern. In this system, the light pen has as at least as many sensors as the number of patterns in the temporal sequence, i.e., twenty-two . In an alternative embodiment, two projectors are used, a first projectordisplays the imperceptible patterns 111, while a second projector 121 displays the underlying perceptible content 122. A third projector, not shown, can interleave a second sequence of patterns so that 3D coordinates of the location information can be obtained in the case where the display surface 610 is curved as shown in Figure 6, and as described in U.S. Patent Application Sn. 10/394,315, "Method and System for Displaying Images on Curved Surfaces," filed by Raskar et al . , on March 21, 2003, incorporated herein by reference.
In yet an another alternative embodiment shown in Figure 5, the display is a mosaic of partially overlapping images 501 to generate a largerpanoramic images 510 as described inU. S . Patent Application 10/394,688, "Self-Configurable Ad-Hoc Projector Cluster, " filedby Raskar et al . , onMarch21, 2003, incorporated herein by reference. In this case, each header sequence 210 can include an identification so that the decoder can differentiate the different sequences of patterns projected by multiple projectors .
It should also be noted, that multiple light pens can be used simultaneously in a multi-user interface. Thus, the invention has advantages over resistive touch screens, which are relatively expensive and can only distinguish a single touched location, at anyone time. Conventional vision-based systems are also more complex to implement and those type of systems have a problem with shadowing and accuracy.
Although the invention has been described by way of examples of preferred embodiments, it is to be understood that various other adaptations andmodifications canbemade within the spirit and scope of the invention. Therefore, it is the object of the appended claims to cover all such variations and modifications as come within the true spirit and scope of the invention.

Claims

1. A method for determining a location on a display surface, comprising: projecting a sequence of patterns onto the display surface, the sequence of patterns having a unique sequence of light intensities for each location of the display surface; sensing intensities of light at an arbitrary location while projecting the sequence of patterns; and decoding the intensities of light to determine coordinates of the arbitrary location.
2. The method of claim 1, in which the calibration patterns are in a form of Gray codes.
3. The method of claim 1, in which the display surface is planar and the coordinates are two-dimensional.
4. The method of claim 1, in which the display surface is quadric and the coordinates are three-dimensional.
5. The method of claim 1, in which the light pen and a projector are on opposite sides of the display surface.
6. The method of claim 1, in which the light pen and a projector are on the same side of the display surface.
7. The method of claim 1, in which the optical sensor is a single photo diode.
8. The method of claim 1 , in which the optical sensor is a camera with a plurality of light sensitive elements.
9. The method of claim 1, in which the light is infrared.
10. The method of claim 1, in which the sequence of patterns is generated as pairs, a second pattern in the pair being an inverse of the first pattern in the pair.
11. Themethodof claiml, inwhich display surface ispixel-based.
12. The method of claim 1, in which the sequence of patterns includes a header sequence of patterns.
13. The method of claim 12, further comprising: detecting the header sequence of patterns to determine a start of the sequence of patterns.
14. The method of claim 12, further comprising: determining a rate of the patterns from the header sequence of patterns.
15. The method of claim 12, further comprising: calibrating the light pen to relative light intensities from the header sequence of patterns.
16. The method of claim 1, in which the sequence of patterns is imperceptible to a human visual system.
17. The method of claim 16, further comprising: displaying perceptible content while displaying the sequence of patterns.
18. The method of claim 12, in which the header sequence includes an identification of the sequence of patterns.
19. An interactive display system, comprising: a display surface; a proj ector configured to display a sequence of patterns onto the display surface, the sequence ofpatterns having a unique sequence of light intensities for each location of the display surface; an optical sensor configured to sense intensities of light at an arbitrary location on the display surface, while the projector displays the sequence of patterns; and means for decoding the intensities of light to determine coordinates of the arbitrary location.
20. The system of claim 19, in which the display surface is pixel-based, and the patterns are Gray codes
PCT/JP2004/016625 2003-11-07 2004-11-02 Method for determining location on display surface and interactive display system WO2005045655A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP2006519005A JP4675889B2 (en) 2003-11-07 2004-11-02 Position detection method, interactive display system, and position detection apparatus
DE602004031140T DE602004031140D1 (en) 2003-11-07 2004-11-02 METHOD FOR DETERMINING THE POSITION ON A DISPLAY SURFACE AND INTERACTIVE DISPLAY SYSTEM
EP04799563A EP1680732B1 (en) 2003-11-07 2004-11-02 Method for determining location on display surface and interactive display system

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/704,098 US7421111B2 (en) 2003-11-07 2003-11-07 Light pen system for pixel-based displays
US10/704,098 2003-11-07

Related Child Applications (1)

Application Number Title Priority Date Filing Date
EP10011898.3A Previously-Filed-Application EP2330484A3 (en) 2003-11-07 2004-11-02 Method for determining location on display surface and interactive display system

Publications (2)

Publication Number Publication Date
WO2005045655A2 true WO2005045655A2 (en) 2005-05-19
WO2005045655A3 WO2005045655A3 (en) 2006-05-18

Family

ID=34552043

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2004/016625 WO2005045655A2 (en) 2003-11-07 2004-11-02 Method for determining location on display surface and interactive display system

Country Status (6)

Country Link
US (1) US7421111B2 (en)
EP (2) EP1680732B1 (en)
JP (2) JP4675889B2 (en)
CN (1) CN100388173C (en)
DE (1) DE602004031140D1 (en)
WO (1) WO2005045655A2 (en)

Families Citing this family (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5356650B2 (en) * 2004-01-15 2013-12-04 テクニオン リサーチ アンド ディベロップメント ファウンデーション リミテッド 3D video scanner
JPWO2007105634A1 (en) * 2006-03-09 2009-07-30 パナソニック株式会社 Wireless IC tag, document medium, and information processing apparatus
JP5317206B2 (en) * 2006-09-21 2013-10-16 トムソン ライセンシング Method and system for 3D model acquisition
JP2008146463A (en) * 2006-12-12 2008-06-26 Mitsubishi Electric Corp Position detection apparatus
US20080170033A1 (en) * 2007-01-15 2008-07-17 International Business Machines Corporation Virtual pointer
US8994757B2 (en) 2007-03-15 2015-03-31 Scalable Display Technologies, Inc. System and method for providing improved display quality by display adjustment and image processing using optical feedback
US20090219262A1 (en) * 2007-12-29 2009-09-03 Microvision, Inc. Active Input Device for a Scanned Beam Display
US8519983B2 (en) * 2007-12-29 2013-08-27 Microvision, Inc. Input device for a scanned beam display
US20090189858A1 (en) * 2008-01-30 2009-07-30 Jeff Lev Gesture Identification Using A Structured Light Pattern
US20090219381A1 (en) * 2008-03-03 2009-09-03 Disney Enterprises, Inc., A Delaware Corporation System and/or method for processing three dimensional images
JP5224973B2 (en) * 2008-08-26 2013-07-03 株式会社ジャパンディスプレイウェスト Information input / output device and information input / output method
US9753558B2 (en) * 2009-01-22 2017-09-05 Texas Instruments Incorporated Pointing system and method
JP4699536B2 (en) * 2009-03-06 2011-06-15 シャープ株式会社 POSITION DETECTION DEVICE, CONTROL METHOD, CONTROL PROGRAM, AND RECORDING MEDIUM
US20100321382A1 (en) 2009-06-18 2010-12-23 Scalable Display Technologies, Inc. System and method for injection of mapping functions
CN101963846B (en) * 2009-07-24 2013-04-24 精工爱普生株式会社 Optical pen
KR101407818B1 (en) * 2009-12-08 2014-06-17 한국전자통신연구원 Apparatus and method for extracting depth image and texture image
CN101751161B (en) * 2009-12-23 2012-10-17 明基电通有限公司 Optical pen, optical pen position decision method and interactive display system
US8217997B2 (en) * 2010-03-16 2012-07-10 Interphase Corporation Interactive display system
CN102243417A (en) 2010-05-10 2011-11-16 中强光电股份有限公司 Optical device and projection system
KR20180117717A (en) 2010-11-15 2018-10-29 스케일러블 디스플레이 테크놀로지스, 인크. System and method for calibrating a display system using manual and semi-automatic techniques
US9639178B2 (en) * 2010-11-19 2017-05-02 Apple Inc. Optical stylus
US8872763B2 (en) * 2011-01-19 2014-10-28 Seiko Epson Corporation Position detecting system and position detecting method
CN102768580B (en) * 2011-05-04 2015-11-25 台达电子工业股份有限公司 For the indicating device of an optical projection system
US9497447B2 (en) 2011-06-15 2016-11-15 Scalable Display Technologies, Inc. System and method for color and intensity calibrating of a display system for practical usage
US9046933B2 (en) 2011-07-19 2015-06-02 Mckesson Financial Holdings Displaying three-dimensional image data
US8963893B2 (en) 2011-08-16 2015-02-24 Plasmability, Llc CRT light pen interface for flat panel displays
US9774989B2 (en) * 2011-09-27 2017-09-26 Sony Interactive Entertainment Inc. Position and rotation of a portable device relative to a television screen
US8963838B2 (en) * 2011-10-25 2015-02-24 Texas Instruments Incorporated Enhanced projected image interface
DE102011086318A1 (en) * 2011-11-14 2013-05-16 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Position determination of an object by detection of a position pattern by optical sensor
CN103197805A (en) * 2012-01-10 2013-07-10 深圳泰山在线科技有限公司 System and method for realizing virtual screen
US8817111B2 (en) 2012-04-19 2014-08-26 Scalable Display Technologies, Inc. System and method of calibrating a display system free of variation in system input resolution
CN102740020B (en) * 2012-05-28 2015-01-21 歌尔声学股份有限公司 Control method of plasma television, Bluetooth touch pen and plasma television
US8922486B2 (en) * 2012-07-24 2014-12-30 Christie Digital Systems Usa, Inc. Method, system and apparatus for determining locations in a projected image
CN103576992B (en) * 2012-08-06 2017-09-15 原相科技股份有限公司 Film and light guide plate and position detection system with positional information
CN102929409B (en) * 2012-10-12 2016-01-13 明基电通有限公司 Telechiric device, display system and display packing
CN103869951A (en) * 2012-12-17 2014-06-18 鸿富锦精密工业(武汉)有限公司 Shadow character input system and method
JP6136264B2 (en) * 2012-12-28 2017-05-31 株式会社リコー Image projection system, projection display device, and program
CN105308503A (en) 2013-03-15 2016-02-03 斯加勒宝展示技术有限公司 System and method for calibrating a display system using a short throw camera
US9524059B2 (en) * 2013-03-15 2016-12-20 Texas Instruments Incorporated Interaction detection using structured light images
JP6286841B2 (en) * 2013-03-18 2018-03-07 セイコーエプソン株式会社 Projector and control method
US9058068B2 (en) * 2013-03-28 2015-06-16 Blackberry Limited Photodetector-based stylus system
CN103336634B (en) * 2013-07-24 2016-04-20 清华大学 Based on touching detection system and the method for adaptive layered structured light
CN104461415B (en) * 2013-09-17 2018-08-10 联想(北京)有限公司 Equipment localization method, device and electronic equipment based on equipment collaboration system
US9347833B2 (en) * 2013-10-10 2016-05-24 Qualcomm Incorporated Infrared touch and hover system using time-sequential measurements
US9207780B2 (en) * 2014-01-27 2015-12-08 Fuji Xerox Co., Ltd. Systems and methods for hiding and finding digital content associated with physical objects via coded lighting
US10318014B2 (en) 2014-01-31 2019-06-11 Hewlett-Packard Development Company, L.P. Optical pointer illumination
EP3141864B1 (en) * 2014-04-30 2019-06-05 Shinano Kenshi Co., Ltd. Measurement device
WO2015166915A1 (en) 2014-04-30 2015-11-05 シナノケンシ株式会社 Measurement device
CN105208181B (en) * 2014-06-11 2018-01-23 联想(北京)有限公司 Information processing method and electronic equipment
JP2016122345A (en) * 2014-12-25 2016-07-07 株式会社リコー Image projection device and interactive input/output system
US10067905B2 (en) 2015-05-26 2018-09-04 Plasmability, Llc Digital interface for manufacturing equipment
CN107979748A (en) * 2016-10-21 2018-05-01 中强光电股份有限公司 Projector, optical projection system and image projecting method
JP7367731B2 (en) 2021-05-12 2023-10-24 セイコーエプソン株式会社 Projection system control method and projection system

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003017076A1 (en) 2001-08-20 2003-02-27 Chi-Lei Kao Input system and method for coordinate and pattern

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2632058A (en) 1946-03-22 1953-03-17 Bell Telephone Labor Inc Pulse code communication
US4357108A (en) * 1980-06-06 1982-11-02 Robotic Vision Systems, Inc. Method for reproducton of object surfaces
US4792209A (en) * 1986-05-30 1988-12-20 Laine Curtis S Projection screen
GB8701206D0 (en) * 1987-01-20 1987-02-25 Hilton C S Apparatus for capturing information in drawing/writing
DE3880847T2 (en) * 1987-01-20 1993-11-18 British Tech Group Method and device for taking information when drawing or writing.
JPH06506080A (en) * 1991-04-03 1994-07-07 ヒューレット・パッカード・カンパニー position sensing device
CA2060564C (en) * 1992-02-06 1996-05-21 Toru Suzuki Wireless input system for computer
US5465121A (en) * 1993-03-31 1995-11-07 International Business Machines Corporation Method and system for compensating for image distortion caused by off-axis image projection
US6005990A (en) * 1996-06-10 1999-12-21 International Business Machines Corporation System for optically scanning images using optical shutter
US6570623B1 (en) * 1999-05-21 2003-05-27 Princeton University Optical blending for multi-projector display wall systems
US6489961B1 (en) * 2000-10-17 2002-12-03 Actuality Systems, Inc. Rasterization of lines in a cylindrical voxel grid
JP2003091366A (en) * 2001-09-17 2003-03-28 Fuji Photo Optical Co Ltd Presentation system
DE60205662T2 (en) * 2001-12-21 2006-06-29 British Telecommunications P.L.C. Apparatus and method for calculating a position of a display
JP4217021B2 (en) * 2002-02-06 2009-01-28 株式会社リコー Coordinate input device
US7148933B2 (en) * 2002-05-03 2006-12-12 Hewlett-Packard Development Company, L.P. Projector having alignment optics and electronics
WO2004001332A1 (en) * 2002-06-19 2003-12-31 Canesta, Inc. System and method for determining 3-d coordinates of a surface using a coded array
US7146036B2 (en) * 2003-02-03 2006-12-05 Hewlett-Packard Development Company, L.P. Multiframe correspondence estimation

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003017076A1 (en) 2001-08-20 2003-02-27 Chi-Lei Kao Input system and method for coordinate and pattern

Also Published As

Publication number Publication date
EP1680732B1 (en) 2011-01-19
DE602004031140D1 (en) 2011-03-03
EP2330484A2 (en) 2011-06-08
US20050099405A1 (en) 2005-05-12
JP2007510966A (en) 2007-04-26
CN1875337A (en) 2006-12-06
JP2011014168A (en) 2011-01-20
EP2330484A3 (en) 2015-07-01
WO2005045655A3 (en) 2006-05-18
CN100388173C (en) 2008-05-14
JP4745454B2 (en) 2011-08-10
JP4675889B2 (en) 2011-04-27
US7421111B2 (en) 2008-09-02
EP1680732A2 (en) 2006-07-19

Similar Documents

Publication Publication Date Title
US7421111B2 (en) Light pen system for pixel-based displays
US10085002B2 (en) RGB-IR sensor, and method and apparatus for obtaining 3D image by using same
US11330243B2 (en) System and method for 3D scanning
JP4822643B2 (en) Computer presentation system and method with optical tracking of a wireless pointer
US7973779B2 (en) Detecting ambient light levels in a vision system
EP2026170A1 (en) Position detecting device
MX2011008489A (en) Touch pointers disambiguation by active display feedback.
JP2000357055A (en) Method and device for correcting projection image and machine readable medium
US11056028B2 (en) Method for detecting luminance uniformity of screen, storage medium, and electronic device
TW200939096A (en) Display apparatus and image pickup apparatus
CN102455779A (en) Information processing device and method
KR20170057110A (en) Image apparatus and operation method thereof
JP2015506006A (en) Method for locating an object by detecting a position pattern with an optical sensor
US20120206416A1 (en) Interactive Display
CN101542550A (en) Remote control pointing technology
Park et al. Subjective evaluation on visual perceptibility of embedding complementary patterns for nonintrusive projection-based augmented reality
CN112433640B (en) Automatic calibration interactive projection system of multiple image sensors and implementation method thereof
JP7347205B2 (en) Projection system control method, projection system and control program
TW201322071A (en) Display apparatus and touch sensing method thereof
US9733762B2 (en) Optical display and control element and method of optically determining a position
Chrásková et al. An automatic 3D tracking system with a PC and a single TV camera
JP2012515358A (en) Optical scanning display with high edge reproducibility
CN105843455B (en) Method for setting effective pixel of image sensing array
KR20170124134A (en) Monitor Drive Method for Detecting a Change of the Position of the Optical Pen Mouse Operated on a Display Screen

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200480032382.0

Country of ref document: CN

AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2004799563

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2006519005

Country of ref document: JP

WWP Wipo information: published in national office

Ref document number: 2004799563

Country of ref document: EP