WO2010094065A1 - Augmented reality system and method - Google Patents

Augmented reality system and method Download PDF

Info

Publication number
WO2010094065A1
WO2010094065A1 PCT/AU2010/000173 AU2010000173W WO2010094065A1 WO 2010094065 A1 WO2010094065 A1 WO 2010094065A1 AU 2010000173 W AU2010000173 W AU 2010000173W WO 2010094065 A1 WO2010094065 A1 WO 2010094065A1
Authority
WO
WIPO (PCT)
Prior art keywords
marker
augmented reality
view
field
detected
Prior art date
Application number
PCT/AU2010/000173
Other languages
French (fr)
Inventor
Adrian Risch
Dean Farmer
Original Assignee
Jumbuck Entertainment Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from AU2009900669A external-priority patent/AU2009900669A0/en
Application filed by Jumbuck Entertainment Limited filed Critical Jumbuck Entertainment Limited
Publication of WO2010094065A1 publication Critical patent/WO2010094065A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • G06F3/0425Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means using a single imaging device like a video camera for tracking the absolute position of a single or a plurality of objects with respect to an imaged reference surface, e.g. video camera imaging a display or a projection screen, a table or a wall surface, on which a computer generated image is displayed or projected
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30204Marker

Definitions

  • the invention relates to an augmented reality system and method.
  • the present invention relates to an augmented reality system in which actions are taken in relation to computer generated objects in response to the failure to detect a marker associated with that object.
  • Augmented reality (“AR”) systems are computer generated systems that process both real world and computer-generated data.
  • AR systems are able to blend (or "augment") real world video footage - captured by digital video cameras - with computer-generated graphical objects, that appear in real time in the captured video footage.
  • a classic example of an AR system is the "headup display" used in fighter aircraft, in which information about the aircraft's position and speed or regarding an enemy aircraft is displayed to the pilot superimposed over the real image of the pilot's view.
  • the additional information is sourced from various locations including radar and data from the aircraft's object recognition systems. This overlaying of computer generated information onto real world information augments the pilot's knowledge and understanding of the field of view.
  • Some AR systems utilise objects in the video footage along with information about the relative angle and position of the camera used to capture the footage, in order to overlay graphical objects in a convenient location. For example, lines used on a tennis court or football field are used as references to overlay graphical objects on the video footage broadcast to viewers.
  • Other AR systems rely entirely on the location of a detectable object such as a barcode or fiducial marker (see below) within the camera frame.
  • print advertisements sometimes include a barcode or fiducial marker, which, when viewed on a computer monitor using a PC camera and associated decoding software, displays a three-dimensional model of an object such as a new car. As the advertisement and associated barcode or fiducial marker rotates or moves, the corresponding computer graphic object of the car also moves, creating an impression of a real car that is viewable by a potential buyer from different angles.
  • AR thus provides a means of enhancing the physical world beyond normal human experience.
  • markers that are used by motion tracking algorithms as measures and as points of reference.
  • Markers - which may be considered as AR objects in themselves - are known in the art as “fiducial markers” or “fiducials”.
  • Other types of markers used in AR systems include images and special codes, such as BCH codes and barcodes.
  • the graphical object associated with the relevant real world object is rendered at the current position of the marker, or at another point as determined by the underlying AR software. If a marker is obscured from view, such as by an errant finger in the video camera, or by another object appearing in the captured video scene, conventional current AR systems cease rendering the object. Accordingly, the object disappears from the AR display (which may be a computer monitor, video camera, mobile phone, headgear etc.) and is only rendered again when the marker is detected at a later time. Objects can then flicker in and out of existence on the display, producing an unrealistic and distracting effect.
  • the AR display which may be a computer monitor, video camera, mobile phone, headgear etc.
  • the object continues to be rendered onscreen even after the marker fails to be detected, in order to create a more realistic scene.
  • the marker is redetected, possibly in a new location, the object is re-rendered at its new location
  • this implementation is problematic in the scenario where the marker has moved (possibly permanently) out of the camera's field of view, as opposed to where the marker continues to be within the field of view but nevertheless is not detected.
  • Current systems continue to render the object at its last known location, which tends to result in the creation of a somewhat unrealistic scene. Objects can appear to jump from point to point on a monitor, or remain stationary when in fact the marker may have moved permanently out of the field of view.
  • the present invention aims to address or ameliorate this deficiency in current AR systems.
  • a method for implementing an augmented reality system comprising one or more graphical objects, each having an associated marker, the method including the steps of: displaying a graphical object when its associated marker is detected by a detection means; and upon a failure to detect the marker, ceasing the display of the graphical object if the marker was last detected in a selected region of the field of view of the detection means, or continuing the display of the graphical object if the marker was last detected in another part of the field of view of the detection means apart from the selected region.
  • the present invention predicates the decision whether to continue or cease displaying a graphical object in an augmented reality system on the basis of the last known location of the object's associated marker. It has been realised that markers may fail to be detected for a number of different reasons and that the last known location of the marker can provide valuable information regarding why the marker was not detected.
  • the selected region is an area within the field of view, in which a failure to detect the marker can be attributed to the marker moving out of the field of view.
  • graphical objects continue to be displayed when it is more likely that the marker was obscured by another object in the field of view (such as a finger, hand or pencil) rather than the detection-failure resulting from the marker being physically moved outside of the field of view.
  • display of the graphical object is ceased when it is more likely that the detection- failure was due to the marker moving outside of the field of view.
  • Realistic augmented reality scenes can be created through use of this embodiment of the invention.
  • the area of said selected region is the periphery of the field of view.
  • the method includes the step of periodically flagging each marker with a value indicating whether the marker was detected in the selected region.
  • the value is examined and the display of the associated object either ceased or continued accordingly.
  • the method may include the step of transitioning the graphical object from a previously detected position to a new detected position in the display.
  • Transitioning may include applying an animation to the object, such that the object is continuously displayed between the previously detected position and the new position. Transitioning may also include moving the object a proportion of the distance between the previously detected position and the new detected position at a set distance per time unit.
  • Animations are typically applied on the basis of the object and may include walking, running, driving and the like.
  • an augmented reality system comprising: a processor; a camera coupled to the processor for capturing video within a field of view; a display coupled to the processor for displaying the captured video; and an augmented reality application adapted for execution on the processor, the application having: a detection routine for detecting one or more markers in the captured video and associating the marker with a graphical object displayable on the display; and a display routine configured to, upon a failure to detect a marker, cease display of a graphical object if its associated marker was last detected in a selected region of the field of view, or continue display of the graphical object, if its associated marker was last detected in another part of the field of view apart from the selected region.
  • the selected region is an area within the field of view, in which a failure to detect the marker can be attributed to the marker moving out of the field of view.
  • the area of said selected region is the periphery of the field of view.
  • the augmented reality application includes a flagging routine for periodically flagging the or each marker with a value indicating whether the marker was detected in the selected region. Upon a failure to detect a marker, the value is examined and the display of the associated object either ceased or continued accordingly. '
  • the augmented reality application includes a transitioning routine for transitioning objects from a previously detected position to a new detected position.
  • the transitioning routine may be configured to apply an animation to the object, such that the object is continuously displayed between the previously detected position and the new position. Transitioning may also include moving the object a proportion of the distance between the previously detected position and the new detected position at a set distance per time unit. Animations are typically applied on the basis of the object and may include walking, running, driving and the like.
  • a software product stored on a computer readable medium, the software product including computer program code which, when executed by a processor, is configured to execute the method according to the first aspect of the invention or implement the augmented reality system according to the second aspect of the invention.
  • Figures 1 to 4 are screen shots of an AR system implemented in accordance with the present invention, showing a fiducial marker with its associated graphical object;
  • Figure 5 is a flow chart of the method steps for implementing an AR system in accordance with an embodiment of the present invention.
  • Figures 6 to 9 are schematics, illustrating alternative implementations of a peripheral region and an interior region of a field of view in an AR system in accordance with an embodiment of the present invention.
  • FIGS. 10 to 13 are schematics, illustrating an AR system in accordance with an embodiment of the present invention
  • an AR system blends real world video footage captured by a detection means, such as a digital video camera, with computer-generated graphical objects.
  • a detection means such as a digital video camera
  • a number of software platforms are currently available for developing AR systems, such as the ARToolkit from Canterbury University, Wales, New Zealand and Studierstube from Graz University, Austria. These platforms provide APIs and library routines implementing basic AR functionalities, including image detection and fiducial marker recognition and tracking.
  • markers may be images or special codes such as BCH codes or bar codes.
  • FIG 1 a scene created in an AR system and displayed on a PC screen is illustrated.
  • a fiducial marker 10 is located on a surface in front of the PC and is detected by a digital video camera (not shown) attached to the PC. The camera captures video footage only within a bounded field of view 13.
  • Field of view 13 is illustrated in Figures 1 to 4 as a lighter rectangular area within a larger, blackened rectangular area.
  • Fiducial marker 10 has an associated graphical object 12 that is generated by the software platform running on the PC and displayed on the PC screen blended with the video footage captured by the PC camera. Object 12 is displayed when its associated fiducial marker 10 is detected by the PC camera.
  • a fiducial marker may fail to be detected when it is obscured by another object in field of view 13 of the video camera. This is illustrated in Figure 2 by a hand 14 covering fiducial marker 10.
  • a fiducial marker may also fail to be detected when it moves completely out of field of view 13. In the former instance, it is preferable to continue displaying an associated graphical object, because the marker remains within the field of view and the disappearance of the marker is more likely to be temporary. Ceasing the display of an object each and every time its marker disappears from view, possibly for a limited time period, results in a somewhat unrealistic AR scene and an unsatisfactory user experience.
  • a first region 18 is the area defined within boundary 16 and a second region 20 is the area defined outside boundary 16 up to the edge of field of view 13.
  • Region 20 is marked as 'Dead zone' in Figures 1 to 4 and is the periphery of field of view 13.
  • second region 20 can be considered as a 'warning area' for markers and their associated graphical objects.
  • the inventors have realised that when a marker is located in second region 20 and then subsequently fails to be detected, it is more likely that the detection-failure is due to the marker leaving field of view 13 than the marker being obscured by another object in field of view 13.
  • FIG. 5 A flow chart of a method used to implement the AR system functionality is shown in Figure 5. The method steps are executed by a software product built on top of a suitable AR software platform, such as those discussed above.
  • all of the markers present in the AR scene are detected, by way of suitable image detection/marker tracking algorithms.
  • a decision is made as to whether a particular marker is detected. Markers are detected when the tracking algorithm identifies at least one marker within the camera's view angle. Where a marker is detected, a pose estimation algorithm is deployed to determine the marker's position and orientation. This assists in determining whether the marker is in an acceptable (see below) area of the camera's view.
  • Markers according to the invention are created as data structures having a number of attributes. Each attribute may be assigned a selected value.
  • an important marker attribute relates to whether the marker is a "removable" marker or a "non-removable” marker.
  • An attribute is a variable whose value is varied in response to events occurring within the AR system (see below) .
  • the object associated with the marker is not rendered on the AR display (step 36).
  • the associated objected is rendered in the display in its last known position (step 38).
  • Rendering a graphic object associated with a non-removable marker in the event of not detecting that marker, maintains a continuous and consistent rendering of the object. Rendered objects do not "jump" throughout the display and are not constantly rendered and deleted in an attempt to follow a fast moving and disappearing marker, as occurred in prior art AR systems.
  • step 30 The method returns to step 30, where markers are again detected, after either step 36 (where the object is not displayed) or step 38 (where the object is displayed in its last known position) .
  • 'non-valid' locations can be considered dead zones, including locations at a depth considered prohibitive for correct tracking.
  • the relevant attribute value of the marker (at step 42) is set to 'removable'. It will be realised that the marker will not be displayed if it subsequently fails to be detected during later iterations of ste ⁇ s 32 to 38.
  • the marker's attribute value (at step 44) is set to 'non-removable'.
  • markers with 'non-removable' attributes continue to be displayed notwithstanding a failure to detect the marker.
  • the associated object is animated and displayed at its new location.
  • Objects are animated to their new position based on an algorithmic interpolation between the old and new location and/or orientation. Action is taken to ensure a seamless transition of the object to its new location. Usually, this is in the form of applying an appropriate animation (walking, running, driving, etc) from the old location to the new location.
  • step 50 the associated object is again rendered at its current position.
  • the most recent rendered position of the object is recorded for further processing.
  • 'exit animations' can be included, by which (if an object is to be removed from the display), it leaves the scene in a suitable manner (i.e. a graceful departure, running, driving, etc).
  • a similar approach can be taken for introduction of objects newly detected, with objects animated from off-screen to a position corresponding to the first detected position of the marker.
  • the process returns to step 30 for further detection of markers in the AR system.
  • a dead zone for use in accordance with the invention can be marked using: pixels; or percentage width of the screen
  • a pixel marking does not vary with the screen resolution of the display, i.e. it is always a set number of pixels from the edge of the screen (For example 5 ⁇ x). If the screen resolution is 320x240, then the renderable area would be 310x230. ( Figure 6). Alternatively, if the screen resolution is 640x480, then the render able zone would be 630x470. ( Figure 7)
  • a percentage width-marking does vary with the screen resolution i.e. is a relative width from the edge of the screen (For example 10%). If the screen resolution is 320x240, then the render able zone would be 288x216 ( Figure 8) If the screen resolution is 640x480, then the render able zone would be 576x432. ( Figure 9).

Abstract

The present invention relates to a method for implementing an augmented reality system, the augmented reality system comprising one or more graphical objects (12), each having an associated marker (10). The method includes the steps of: displaying a graphical object (12) when its associated marker (10) is detected by a detection means; and upon a failure to detect the marker (34), ceasing the display (36) of the graphical object (12) if the marker (10) was last detected in a selected region (20) of the field of view of the detection means, or continuing (38) the display of the graphical object (12) if the marker (10) was last detected in another part (18) of the field of view of the detection means apart from the selected region (20).

Description

Augmented reality system and method Field of the Invention
The invention relates to an augmented reality system and method. In particular, the present invention relates to an augmented reality system in which actions are taken in relation to computer generated objects in response to the failure to detect a marker associated with that object.
Background of the Invention
In this specification, where a document, act or item of knowledge is referred to or discussed, this reference or discussion is not an admission that the document, act or item of knowledge or any combination thereof was at the priority date:
(i) part of common general knowledge; or
(ii) known to be relevant to an attempt to solve any problem with which this specification is concerned.
Augmented reality ("AR") systems are computer generated systems that process both real world and computer-generated data. In particular, AR systems are able to blend (or "augment") real world video footage - captured by digital video cameras - with computer-generated graphical objects, that appear in real time in the captured video footage.
A classic example of an AR system is the "headup display" used in fighter aircraft, in which information about the aircraft's position and speed or regarding an enemy aircraft is displayed to the pilot superimposed over the real image of the pilot's view. The additional information is sourced from various locations including radar and data from the aircraft's object recognition systems. This overlaying of computer generated information onto real world information augments the pilot's knowledge and understanding of the field of view.
Some AR systems utilise objects in the video footage along with information about the relative angle and position of the camera used to capture the footage, in order to overlay graphical objects in a convenient location. For example, lines used on a tennis court or football field are used as references to overlay graphical objects on the video footage broadcast to viewers. Other AR systems rely entirely on the location of a detectable object such as a barcode or fiducial marker (see below) within the camera frame. For example, print advertisements sometimes include a barcode or fiducial marker, which, when viewed on a computer monitor using a PC camera and associated decoding software, displays a three-dimensional model of an object such as a new car. As the advertisement and associated barcode or fiducial marker rotates or moves, the corresponding computer graphic object of the car also moves, creating an impression of a real car that is viewable by a potential buyer from different angles.
AR thus provides a means of enhancing the physical world beyond normal human experience. Within this field there is a huge and growing range of applications, including the areas of education, games and entertainment.
As mentioned above, real world objects appearing in the footage may be tagged with markers, that are used by motion tracking algorithms as measures and as points of reference. Markers - which may be considered as AR objects in themselves - are known in the art as "fiducial markers" or "fiducials". Other types of markers used in AR systems include images and special codes, such as BCH codes and barcodes.
When a marker is detected, the graphical object associated with the relevant real world object is rendered at the current position of the marker, or at another point as determined by the underlying AR software. If a marker is obscured from view, such as by an errant finger in the video camera, or by another object appearing in the captured video scene, conventional current AR systems cease rendering the object. Accordingly, the object disappears from the AR display (which may be a computer monitor, video camera, mobile phone, headgear etc.) and is only rendered again when the marker is detected at a later time. Objects can then flicker in and out of existence on the display, producing an unrealistic and distracting effect.
In other instances, the object continues to be rendered onscreen even after the marker fails to be detected, in order to create a more realistic scene. When the marker is redetected, possibly in a new location, the object is re-rendered at its new location However, this implementation is problematic in the scenario where the marker has moved (possibly permanently) out of the camera's field of view, as opposed to where the marker continues to be within the field of view but nevertheless is not detected. Current systems continue to render the object at its last known location, which tends to result in the creation of a somewhat unrealistic scene. Objects can appear to jump from point to point on a monitor, or remain stationary when in fact the marker may have moved permanently out of the field of view.
The present invention aims to address or ameliorate this deficiency in current AR systems.
Summary of the Invention
According to a first aspect of the present invention there is provided a method for implementing an augmented reality system, the augmented reality system comprising one or more graphical objects, each having an associated marker, the method including the steps of: displaying a graphical object when its associated marker is detected by a detection means; and upon a failure to detect the marker, ceasing the display of the graphical object if the marker was last detected in a selected region of the field of view of the detection means, or continuing the display of the graphical object if the marker was last detected in another part of the field of view of the detection means apart from the selected region.
The present invention predicates the decision whether to continue or cease displaying a graphical object in an augmented reality system on the basis of the last known location of the object's associated marker. It has been realised that markers may fail to be detected for a number of different reasons and that the last known location of the marker can provide valuable information regarding why the marker was not detected.
Preferably, the selected region is an area within the field of view, in which a failure to detect the marker can be attributed to the marker moving out of the field of view. According to this embodiment, graphical objects continue to be displayed when it is more likely that the marker was obscured by another object in the field of view (such as a finger, hand or pencil) rather than the detection-failure resulting from the marker being physically moved outside of the field of view. Conversely, display of the graphical object is ceased when it is more likely that the detection- failure was due to the marker moving outside of the field of view. Realistic augmented reality scenes can be created through use of this embodiment of the invention.
Typically, the area of said selected region is the periphery of the field of view. When a marker fails to be detected and was last known to be located in the periphery of the field of view, it is likely that the detection-failure is a result of the marker moving outside of the field of view, rather than a result of the marker being obscured by another object in the field of view.
Preferably, the method includes the step of periodically flagging each marker with a value indicating whether the marker was detected in the selected region. According to this embodiment, upon a failure to detect a marker, the value is examined and the display of the associated object either ceased or continued accordingly.
Optionally, the method may include the step of transitioning the graphical object from a previously detected position to a new detected position in the display.
Transitioning may include applying an animation to the object, such that the object is continuously displayed between the previously detected position and the new position. Transitioning may also include moving the object a proportion of the distance between the previously detected position and the new detected position at a set distance per time unit.
Animations are typically applied on the basis of the object and may include walking, running, driving and the like.
According to a second aspect of the present invention there is provided an augmented reality system comprising: a processor; a camera coupled to the processor for capturing video within a field of view; a display coupled to the processor for displaying the captured video; and an augmented reality application adapted for execution on the processor, the application having: a detection routine for detecting one or more markers in the captured video and associating the marker with a graphical object displayable on the display; and a display routine configured to, upon a failure to detect a marker, cease display of a graphical object if its associated marker was last detected in a selected region of the field of view, or continue display of the graphical object, if its associated marker was last detected in another part of the field of view apart from the selected region.
Preferably, the selected region is an area within the field of view, in which a failure to detect the marker can be attributed to the marker moving out of the field of view.
Typically, the area of said selected region is the periphery of the field of view.
Preferably, the augmented reality application includes a flagging routine for periodically flagging the or each marker with a value indicating whether the marker was detected in the selected region. Upon a failure to detect a marker, the value is examined and the display of the associated object either ceased or continued accordingly.'
Optionally, the augmented reality application includes a transitioning routine for transitioning objects from a previously detected position to a new detected position.
The transitioning routine may be configured to apply an animation to the object, such that the object is continuously displayed between the previously detected position and the new position. Transitioning may also include moving the object a proportion of the distance between the previously detected position and the new detected position at a set distance per time unit. Animations are typically applied on the basis of the object and may include walking, running, driving and the like.
According to a third aspect of the invention there is provided a software product stored on a computer readable medium, the software product including computer program code which, when executed by a processor, is configured to execute the method according to the first aspect of the invention or implement the augmented reality system according to the second aspect of the invention.
Brief Description of the Drawings
An embodiment of the invention will now be described with reference to the accompanying drawings in which:
Figures 1 to 4 are screen shots of an AR system implemented in accordance with the present invention, showing a fiducial marker with its associated graphical object;
Figure 5 is a flow chart of the method steps for implementing an AR system in accordance with an embodiment of the present invention;
Figures 6 to 9 are schematics, illustrating alternative implementations of a peripheral region and an interior region of a field of view in an AR system in accordance with an embodiment of the present invention; and
Figures 10 to 13 are schematics, illustrating an AR system in accordance with an embodiment of the present invention
Detailed Description of the Drawings
The present invention is implemented within the context of an augmented reality (AR) system. As discussed above in the background to the invention, an AR system blends real world video footage captured by a detection means, such as a digital video camera, with computer-generated graphical objects. A number of software platforms are currently available for developing AR systems, such as the ARToolkit from Canterbury University, Christchurch, New Zealand and Studierstube from Graz University, Austria. These platforms provide APIs and library routines implementing basic AR functionalities, including image detection and fiducial marker recognition and tracking. As discussed above, markers may be images or special codes such as BCH codes or bar codes.
Turning to Figure 1, a scene created in an AR system and displayed on a PC screen is illustrated. A fiducial marker 10 is located on a surface in front of the PC and is detected by a digital video camera (not shown) attached to the PC. The camera captures video footage only within a bounded field of view 13. Field of view 13 is illustrated in Figures 1 to 4 as a lighter rectangular area within a larger, blackened rectangular area.
Fiducial marker 10 has an associated graphical object 12 that is generated by the software platform running on the PC and displayed on the PC screen blended with the video footage captured by the PC camera. Object 12 is displayed when its associated fiducial marker 10 is detected by the PC camera.
A fiducial marker may fail to be detected when it is obscured by another object in field of view 13 of the video camera. This is illustrated in Figure 2 by a hand 14 covering fiducial marker 10.
A fiducial marker may also fail to be detected when it moves completely out of field of view 13. In the former instance, it is preferable to continue displaying an associated graphical object, because the marker remains within the field of view and the disappearance of the marker is more likely to be temporary. Ceasing the display of an object each and every time its marker disappears from view, possibly for a limited time period, results in a somewhat unrealistic AR scene and an unsatisfactory user experience.
On the other hand, it is preferable to cease displaying a graphical object when its associated marker has moved completely out of the field of view. Continuing to display an object, possibly at the last known position of its associated marker, after the marker has moved out of the field of view, also results in an unrealistic AR scene and unsatisfactory user experience.
It is therefore apparent that different actions must be taken with regard to either continuing or ceasing the display of a graphical object upon the failure to detect the object's associated marker, depending on the reason for the failure to detect the marker. To investigate the reason for a failure to detect a marker, the inventors have divided field of view 13 by way of a rectangular boundary 16 into two different regions. A first region 18 is the area defined within boundary 16 and a second region 20 is the area defined outside boundary 16 up to the edge of field of view 13. Region 20 is marked as 'Dead zone' in Figures 1 to 4 and is the periphery of field of view 13.
As discussed in further detail below, second region 20 can be considered as a 'warning area' for markers and their associated graphical objects. In particular, the inventors have realised that when a marker is located in second region 20 and then subsequently fails to be detected, it is more likely that the detection-failure is due to the marker leaving field of view 13 than the marker being obscured by another object in field of view 13.
As illustrated in Figure 2, when marker 10 is located in region 18 and subsequently fails to be detected, the detection-failure is likely to result from marker 10 being obscured by another object in field of view 13, and so associated object 12 continues to be displayed notwithstanding the failure to detect marker 10.
Alternatively, as illustrated in Figures 3 and 4, when marker 10 is located in second region 20 and subsequently fails to be detected, the detection failure is likely to result from marker 10 moving out of field of view 13, and so the display of associated object 12 is ceased.
A flow chart of a method used to implement the AR system functionality is shown in Figure 5. The method steps are executed by a software product built on top of a suitable AR software platform, such as those discussed above. At step 30, all of the markers present in the AR scene (the camera's field of view) are detected, by way of suitable image detection/marker tracking algorithms. At step 32, a decision is made as to whether a particular marker is detected. Markers are detected when the tracking algorithm identifies at least one marker within the camera's view angle. Where a marker is detected, a pose estimation algorithm is deployed to determine the marker's position and orientation. This assists in determining whether the marker is in an acceptable (see below) area of the camera's view.
Markers according to the invention are created as data structures having a number of attributes. Each attribute may be assigned a selected value. In particular, an important marker attribute relates to whether the marker is a "removable" marker or a "non-removable" marker. An attribute is a variable whose value is varied in response to events occurring within the AR system (see below) .
At step 34, in the event that a marker is not detected, a determination is made as to whether the marker is removable. This is implemented simply by querying the relevant attribute value of the marker.
If the marker is set to removable, the object associated with the marker is not rendered on the AR display (step 36).
Alternatively, if the marker attribute is not set to removable (i.e. the marker is "non-removable"), the associated objected is rendered in the display in its last known position (step 38).
Rendering a graphic object associated with a non-removable marker, in the event of not detecting that marker, maintains a continuous and consistent rendering of the object. Rendered objects do not "jump" throughout the display and are not constantly rendered and deleted in an attempt to follow a fast moving and disappearing marker, as occurred in prior art AR systems.
The method returns to step 30, where markers are again detected, after either step 36 (where the object is not displayed) or step 38 (where the object is displayed in its last known position) .
At step 40, in the event that a marker is detected, a determination is made as to whether the object is in region 20 (i.e the "dead zone"). As discussed above, when markers move into the dead zone and subsequently fail to be detected, it is likely that the marker has moved out of field of view 13.
Other 'non-valid' locations can be considered dead zones, including locations at a depth considered prohibitive for correct tracking. In the event that a marker is detected in a dead zone, the relevant attribute value of the marker (at step 42) is set to 'removable'. It will be realised that the marker will not be displayed if it subsequently fails to be detected during later iterations of steρs 32 to 38. Alternatively, where a marker is detected outside the dead zone (i.e in region 18), the marker's attribute value (at step 44) is set to 'non-removable'. Similarly, in later iterations of steps 32 to 38, markers with 'non-removable' attributes continue to be displayed notwithstanding a failure to detect the marker.
In either case (i.e. irrespective of whether the marker is either inside or outside the dead zone) a decision is made at step 46 as to whether the marker is detected in a location different from the location in which it was previously rendered.
If the marker is detected at a different position then, at step 48, the associated object is animated and displayed at its new location. Objects are animated to their new position based on an algorithmic interpolation between the old and new location and/or orientation. Action is taken to ensure a seamless transition of the object to its new location. Usually, this is in the form of applying an appropriate animation (walking, running, driving, etc) from the old location to the new location.
Previously, objects were instantaneously removed and re-rendered in their new location, which, as discussed above, tended to create an unsatisfactory user experience.
Where a marker is detected at its last rendered location, at step 50 the associated object is again rendered at its current position.
At step 52, the most recent rendered position of the object is recorded for further processing.
To increase fluency and believability, 'exit animations' can be included, by which (if an object is to be removed from the display), it leaves the scene in a suitable manner (i.e. a graceful departure, running, driving, etc). A similar approach can be taken for introduction of objects newly detected, with objects animated from off-screen to a position corresponding to the first detected position of the marker. The process returns to step 30 for further detection of markers in the AR system. A dead zone for use in accordance with the invention can be marked using: pixels; or percentage width of the screen
As illustrated in Figures 6 and 7, a pixel marking does not vary with the screen resolution of the display, i.e. it is always a set number of pixels from the edge of the screen (For example 5ρx). If the screen resolution is 320x240, then the renderable area would be 310x230. (Figure 6). Alternatively, if the screen resolution is 640x480, then the render able zone would be 630x470. (Figure 7)
As illustrated in Figures 8 and 9, a percentage width-marking does vary with the screen resolution i.e. is a relative width from the edge of the screen (For example 10%). If the screen resolution is 320x240, then the render able zone would be 288x216 (Figure 8) If the screen resolution is 640x480, then the render able zone would be 576x432. (Figure 9).
An embodiment of the invention is further described with reference to Figures 10 to 13. The terms 'render' and 'display' are intended to be used interchangeably. As discussed above, the invention operates along the lines of:
(a) if a fiducial can be detected, render object;
(b) if the fiducial cannot be detected and its last known location is in the dead zone, then stop rendering the object; and (c) if the fiducial cannot be detected and its last know location does not fall in the dead zone, keep rendering the object on the last known location.
In Figures 10 to 13, "P" represents the fiducial, and the suffixed number represents its position (Pl being the first position, P2 being the second position and so on) . In Figure 10, the objects are rendered at Pl and P2, based on concept "a". In Figure 11, the objects are rendered at Pl and P2, based on concept "a".
In Figure 12 the objects are rendered at Pl, P2, P3, P4 as normal. P5 is not detected and the last known location P4 was in the dead zone, so stop rendering the object, based on concept "b". In Figure 13, the objects are rendered at Pl, P2 as normal. P3 is not detected and the last known location P2 was in the render able area, so continue rendering the object, based on concept "c"
The word 'comprising' and forms of the word 'comprising' as used in this description and in the claims do not limit the invention claimed to exclude any variants or additions.
Modifications and improvements to the invention will be readily apparent to those skilled in the art. Such modifications and improvements are intended to be within the scope of this invention.

Claims

Claims:
1. A method for implpmenting an augmented reality system, the augmented reality system comprising one or more graphical objects, each having an associated marker, the method including the steps of: displaying a graphical object when its associated marker is detected by a detection means; and upon a failure to detect the marker, ceasing the display of the graphical object if the marker was last detected in a selected region of the field of view of the detection means, or continuing the display of the graphical object if the marker was last detected in another part of the field of view of the detection means apart from the selected region.
2. A method according to claim 1, wherein the selected region is an area within the field of view in which a failure to detect the marker can be attributed to the marker moving out of the field of view.
3. A method according to claim 1 or claim 2, wherein the area of said selected region is the periphery of the field of view.
4. A method according to any one of claims 1 to 3, further including the step of periodically flagging each marker with a value indicating whether the marker was detected in the selected region.
5. A method according to any one of claims 1 to 4, further including the step of transitioning the graphical object from a previously detected position to a new detected position.
6. A method according to claim 5, wherein the transitioning step includes applying an animation to the graphical object, such that the object is continuously displayed between the previously detected position and the new position.
7. A method according to claim 5, wherein the transitioning step includes moving the graphical object a proportion of the distance between the previously detected position and the new detected position at a set distance per time unit.
8. A method according to claim 6, wherein the animation is applied on the basis of the object and includes one or more of walking, running, driving and the like.
9. An augmented reality system comprising: a processor; a camera coupled to the processor for capturing video within a field of view; a display coupled to the processor for displaying the captured video; and an augmented reality application adapted for execution on the processor, the application having: a detection routine for detecting one or more markers in the captured video and associating the or each marker with a graphical object displayable on the display; and a display routine configured to, upon a failure to detect a marker, cease display of a graphical object if its associated marker was last detected in a selected region of the field of view, or continue display of the graphical object if its associated marker was last detected in another part of the field of view apart from the selected region.
10. An augmented reality system according to claim 9, wherein the selected region is an area within the field of view in which a failure to detect the marker can be attributed to the marker moving out of the field of view.
11. An augmented reality system according to claim 9 or claim 10, wherein the area of said selected region is the periphery of the field of view.
12. An augmented reality system according to any one of claims 9 to 11, wherein the augmented reality application further includes a flagging routine for periodically flagging the or each marker with a value indicating whether the marker was detected in the selected region.
13. An augmented reality system according to any one of claims 9 to 12, wherein the augmented reality application includes a transitioning routine for transitioning objects from a previously detected position to a new detected position.
14. An augmented reality system according to claim 13, wherein the transitioning routine applies an animation to the object, such that the object is continuously displayed between the previously detected position and the new position.
15. An augmented reality system according to claim 13, wherein the transitioning routine moves the object a proportion of the distance between the previously detected position and the new detected position at a set distance per time unit.
16. An augmented reality system according to claim 14, wherein the transitioning routine applies animations on the basis of the object, including walking, running, driving and the like.
17. A software product stored on a computer readable medium, the software product including computer program code which, when executed by a processor, is configured to execute the method according to any one of claims 1 to 8, or implement the augmented reality system according to any one of claims 9 to 16.
PCT/AU2010/000173 2009-02-17 2010-02-17 Augmented reality system and method WO2010094065A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
AU2009900669 2009-02-17
AU2009900669A AU2009900669A0 (en) 2009-02-17 Augmented reality system and method

Publications (1)

Publication Number Publication Date
WO2010094065A1 true WO2010094065A1 (en) 2010-08-26

Family

ID=42633349

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/AU2010/000173 WO2010094065A1 (en) 2009-02-17 2010-02-17 Augmented reality system and method

Country Status (1)

Country Link
WO (1) WO2010094065A1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103270537A (en) * 2011-10-27 2013-08-28 索尼公司 Image processing apparatus, image processing method, and program
GB2502946A (en) * 2012-04-13 2013-12-18 Blippar Com Ltd Maintaining augmented reality image when marker leaves field of view
US8817047B1 (en) 2013-09-02 2014-08-26 Lg Electronics Inc. Portable device and method of controlling therefor
US20150022551A1 (en) * 2013-07-19 2015-01-22 Lg Electronics Inc. Display device and control method thereof
WO2015030305A1 (en) * 2013-08-28 2015-03-05 Lg Electronics Inc. Portable device displaying augmented reality image and method of controlling therefor
WO2015030303A1 (en) * 2013-08-28 2015-03-05 Lg Electronics Inc. Portable device displaying augmented reality image and method of controlling therefor
EP2876608A4 (en) * 2012-07-23 2016-02-10 Fujitsu Ltd Display control program, display control method, and display control device
US9337926B2 (en) 2011-10-31 2016-05-10 Nokia Technologies Oy Apparatus and method for providing dynamic fiducial markers for devices
US9361733B2 (en) 2013-09-02 2016-06-07 Lg Electronics Inc. Portable device and method of controlling therefor
JP2016184296A (en) * 2015-03-26 2016-10-20 富士通株式会社 Display control method, display control program, and information processing apparatus
US10403046B2 (en) * 2017-10-20 2019-09-03 Raytheon Company Field of view (FOV) and key code limited augmented reality to enforce data capture and transmission compliance
CN114339415A (en) * 2021-12-23 2022-04-12 天翼云科技有限公司 Client video playing method and device, electronic equipment and readable medium
US11696011B2 (en) 2021-10-21 2023-07-04 Raytheon Company Predictive field-of-view (FOV) and cueing to enforce data capture and transmission compliance in real and near real time video
US11700448B1 (en) 2022-04-29 2023-07-11 Raytheon Company Computer/human generation, validation and use of a ground truth map to enforce data capture and transmission compliance in real and near real time video of a local scene
US11792499B2 (en) 2021-10-21 2023-10-17 Raytheon Company Time-delay to enforce data capture and transmission compliance in real and near real time video

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5682209A (en) * 1995-11-13 1997-10-28 Tektronix, Inc. Motion estimation using limited-time early exit with prequalification matrices and a predicted search center
US6064749A (en) * 1996-08-02 2000-05-16 Hirota; Gentaro Hybrid tracking for augmented reality using both camera motion detection and landmark tracking

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5682209A (en) * 1995-11-13 1997-10-28 Tektronix, Inc. Motion estimation using limited-time early exit with prequalification matrices and a predicted search center
US6064749A (en) * 1996-08-02 2000-05-16 Hirota; Gentaro Hybrid tracking for augmented reality using both camera motion detection and landmark tracking

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10453266B2 (en) 2011-10-27 2019-10-22 Sony Corporation Image processing apparatus, image processing method, and program
US10902682B2 (en) 2011-10-27 2021-01-26 Sony Corporation Image processing apparatus, image processing method, and program
US9626806B2 (en) 2011-10-27 2017-04-18 Sony Corporation Image processing apparatus, image processing method, and program
US9704297B2 (en) 2011-10-27 2017-07-11 Sony Corporation Image processing apparatus, image processing method, and program
CN103270537A (en) * 2011-10-27 2013-08-28 索尼公司 Image processing apparatus, image processing method, and program
US10068382B2 (en) 2011-10-27 2018-09-04 Sony Corporation Image processing apparatus, image processing method, and program
EP2638527A4 (en) * 2011-10-27 2015-09-02 Sony Corp Image processing apparatus, image processing method, and program
US11468647B2 (en) 2011-10-27 2022-10-11 Sony Corporation Image processing apparatus, image processing method, and program
US9292974B2 (en) 2011-10-27 2016-03-22 Sony Corporation Image processing apparatus, image processing method, and program
US11941766B2 (en) 2011-10-27 2024-03-26 Sony Group Corporation Image processing apparatus, image processing method, and program
US9337926B2 (en) 2011-10-31 2016-05-10 Nokia Technologies Oy Apparatus and method for providing dynamic fiducial markers for devices
GB2502946A (en) * 2012-04-13 2013-12-18 Blippar Com Ltd Maintaining augmented reality image when marker leaves field of view
EP2876608A4 (en) * 2012-07-23 2016-02-10 Fujitsu Ltd Display control program, display control method, and display control device
US9773335B2 (en) 2012-07-23 2017-09-26 Fujitsu Limited Display control device and method
US20150022551A1 (en) * 2013-07-19 2015-01-22 Lg Electronics Inc. Display device and control method thereof
EP3022629A4 (en) * 2013-07-19 2017-03-08 LG Electronics Inc. Display device and control method thereof
WO2015030305A1 (en) * 2013-08-28 2015-03-05 Lg Electronics Inc. Portable device displaying augmented reality image and method of controlling therefor
WO2015030303A1 (en) * 2013-08-28 2015-03-05 Lg Electronics Inc. Portable device displaying augmented reality image and method of controlling therefor
US9361733B2 (en) 2013-09-02 2016-06-07 Lg Electronics Inc. Portable device and method of controlling therefor
US8817047B1 (en) 2013-09-02 2014-08-26 Lg Electronics Inc. Portable device and method of controlling therefor
JP2016184296A (en) * 2015-03-26 2016-10-20 富士通株式会社 Display control method, display control program, and information processing apparatus
US10403046B2 (en) * 2017-10-20 2019-09-03 Raytheon Company Field of view (FOV) and key code limited augmented reality to enforce data capture and transmission compliance
US11696011B2 (en) 2021-10-21 2023-07-04 Raytheon Company Predictive field-of-view (FOV) and cueing to enforce data capture and transmission compliance in real and near real time video
US11792499B2 (en) 2021-10-21 2023-10-17 Raytheon Company Time-delay to enforce data capture and transmission compliance in real and near real time video
CN114339415A (en) * 2021-12-23 2022-04-12 天翼云科技有限公司 Client video playing method and device, electronic equipment and readable medium
CN114339415B (en) * 2021-12-23 2024-01-02 天翼云科技有限公司 Client video playing method and device, electronic equipment and readable medium
US11700448B1 (en) 2022-04-29 2023-07-11 Raytheon Company Computer/human generation, validation and use of a ground truth map to enforce data capture and transmission compliance in real and near real time video of a local scene

Similar Documents

Publication Publication Date Title
WO2010094065A1 (en) Augmented reality system and method
US11019283B2 (en) Augmenting detected regions in image or video data
US20200035019A1 (en) Method and system for generating an image
US8049750B2 (en) Fading techniques for virtual viewpoint animations
US9477965B2 (en) Apparatus and method for identifying insertion zones in video material and for inserting additional material into the insertion zones
US8466913B2 (en) User interface for accessing virtual viewpoint animations
US8885979B2 (en) Apparatus and associated methodology for analyzing subject motion in images
US8154633B2 (en) Line removal and object detection in an image
AU2020203375A1 (en) Systems and methods for tracking and tagging objects within a broadcast
US20150339842A1 (en) System and Method for Rendering in Accordance with Location of Virtual Objects in Real-Time
CA2907709C (en) Field goal indicator for video presentation
EP1922863B1 (en) System and method for managing the visual effects insertion in a video stream
US20220180570A1 (en) Method and device for displaying data for monitoring event
US6945869B2 (en) Apparatus and method for video based shooting game
WO2009064904A1 (en) 3d textured objects for virtual viewpoint animations
US20080291272A1 (en) Method and system for remote estimation of motion parameters
US20090015678A1 (en) Method and system for automatic pose and trajectory tracking in video
Bebie et al. A Video‐Based 3D‐Reconstruction of Soccer Games
EP3836081A1 (en) Data processing method and apparatus
JP2001178963A5 (en)
DK201870226A1 (en) A method and apparatus for user interaction with a video stream
CN114302234B (en) Quick packaging method for air skills
EP4261788A1 (en) Image processing apparatus, image processing method, and program
Lai et al. Tennis video enrichment with content layer separation and real-time rendering in sprite plane
Martín et al. Automatic players detection and tracking in multi-camera tennis videos

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10743321

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10743321

Country of ref document: EP

Kind code of ref document: A1