US20110207089A1 - Firearm training systems and methods of using the same - Google Patents

Firearm training systems and methods of using the same Download PDF

Info

Publication number
US20110207089A1
US20110207089A1 US13/035,321 US201113035321A US2011207089A1 US 20110207089 A1 US20110207089 A1 US 20110207089A1 US 201113035321 A US201113035321 A US 201113035321A US 2011207089 A1 US2011207089 A1 US 2011207089A1
Authority
US
United States
Prior art keywords
firearm
display
tracking
controller
scope
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/035,321
Inventor
David Alfred A. Lagettie
Paige Tedford Manard
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US13/035,321 priority Critical patent/US20110207089A1/en
Publication of US20110207089A1 publication Critical patent/US20110207089A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • FMECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
    • F41WEAPONS
    • F41GWEAPON SIGHTS; AIMING
    • F41G3/00Aiming or laying means
    • F41G3/26Teaching or practice apparatus for gun-aiming or gun-laying
    • F41G3/2605Teaching or practice apparatus for gun-aiming or gun-laying using a view recording device cosighted with the gun
    • F41G3/2611Teaching or practice apparatus for gun-aiming or gun-laying using a view recording device cosighted with the gun coacting with a TV-monitor
    • FMECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
    • F41WEAPONS
    • F41GWEAPON SIGHTS; AIMING
    • F41G3/00Aiming or laying means
    • F41G3/26Teaching or practice apparatus for gun-aiming or gun-laying
    • F41G3/2616Teaching or practice apparatus for gun-aiming or gun-laying using a light emitting device
    • F41G3/2622Teaching or practice apparatus for gun-aiming or gun-laying using a light emitting device for simulating the firing of a gun or the trajectory of a projectile
    • F41G3/2627Cooperating with a motion picture projector

Definitions

  • aspects of the invention relate generally to firearm training systems and, more particularly, to firearm training systems that utilize a simulated virtual environment.
  • Embodiments of the invention may include firearm training systems and methods of using the same.
  • a firearm training system may be provided.
  • the firearm training system may include a firearm, a scope, a tracking system, and at least one controller.
  • the scope may be mounted to the firearm, and the scope may include a display.
  • the tracking system may be configured to collect information associated with a position and an orientation of the firearm.
  • the at least one controller may be configured to (i) determine, based at least in part upon the information collected by the tracking component, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.
  • a system for modifying a firearm for use in a simulation environment may be provided.
  • the system may include a display, a tracking component, and at least one controller.
  • the display may be configured to be attached to a scope associated with the firearm.
  • the tracking component may be configured to be attached to the firearm, and the tracking component may be further configured to collect information associated with a position and an orientation of the firearm.
  • the at least one controller may be configured to (i) determine, based at least in part upon the information collected by the tracking system, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.
  • a firearm scope may be provided.
  • the firearm scope may include a housing, a display, and a video receiver.
  • the housing may be configured to be mounted onto a firearm.
  • the display may be mounted at least partially within the housing.
  • the video receiver may be in communication with the display and configured to output video to be rendered by the display.
  • the video receiver may receive the video from at least one controller configured to (i) determine a viewpoint of the firearm, (ii) adjust a position of the firearm within a virtual environment, (iii) determine the video to be rendered, and (iv) direct communication of the video to the video receiver.
  • a method for providing a firearm simulation may be provided. Motion of a firearm may be tracked via a tracking component attached to the firearm. Based upon information collected by the tracking component, a viewpoint of the firearm may be determined by a controller, such as a controller including one or more computers. Based at least in part upon the determined viewpoint, the controller may determine or identify image data to be rendered by a display associated with a scope mounted onto the firearm. The determined image data may be output by the controller for receipt by the display, and the image data may be presented by the display.
  • a controller such as a controller including one or more computers.
  • the controller may determine or identify image data to be rendered by a display associated with a scope mounted onto the firearm. The determined image data may be output by the controller for receipt by the display, and the image data may be presented by the display.
  • FIG. 1 illustrates a block diagram of an example firearm training system that may be utilized in accordance with various embodiments of the invention.
  • FIG. 2A is an illustration of an actual firearm that has been modified to include attached simulation components.
  • FIG. 2B is an exploded illustration of the firearm shown in FIG. 2A .
  • FIG. 3 illustrates a block diagram of an example scope that may be mounted on a firearm in accordance with various embodiments of the invention.
  • FIG. 4 is an illustration of an actual firearm onto which a replica scope has been mounted to facilitate use of the firearm within a simulation environment.
  • FIG. 5 is an illustration of a firearm into which simulation components have been integrated.
  • FIG. 6 is a diagram of a simulation environment that may be utilized in accordance with various embodiments of the invention.
  • FIG. 7 is a flow diagram of an example method for presenting image data via a display associated with a scope mounted onto a firearm, according to an example embodiment of the invention.
  • FIG. 8 is a flow diagram of an example method for determining image data to be presented via a display associated with a scope mounted onto a firearm, according to an example embodiment of the invention.
  • Embodiments of the invention may include firearm training systems, apparatus, and methods for using the same.
  • a scope may be mounted to a firearm, and a display may be associated with the scope.
  • the display may be integrated into, incorporated into, or at least partially situated within the scope.
  • the display may be attached to the scope.
  • suitable displays may be utilized as desired in various embodiments of the invention, such as a microdisplay, a liquid crystal display (“LCD”), a light emitting diode (“LED”) display, an organic light emitting diode (“OLED”) display, or other suitable display.
  • the display may facilitate the presentation of image data, such as video, to a user of the firearm.
  • one or more displays external to the firearm may be utilized to present a virtual environment to a user.
  • a virtual environment e.g., projection screens, etc.
  • an enlarged representation of a portion of the virtual environment may be presented to the user via the display.
  • At least one tracking system may be associated with the firearm.
  • the tracking system may be configured to track motion, position, and/or orientation data associated with the firearm.
  • a wide variety of suitable tracking systems may be utilized as desired in various embodiments of the invention, including but not limited to, a laser tracking system, an electromagnetic tracking system, a magnetic tracking system, an optical tracking system, an infrared tracking system, an ultrasonic tracking system, an inertia tracking system, a mechanical tracking system, a Global Positioning System (“GPS”) tracking system, or a combination of tracking systems.
  • GPS Global Positioning System
  • motion and/or movement of the firearm and/or scope may be tracked along any number of axes and/or in a wide variety of different directions. For example, movement may be tracked along three axes (e.g., the x, y, and z axes) with anywhere from two to six degrees of freedom.
  • Information collected by the tracking system may be utilized to determine a viewpoint associated with the firearm.
  • information collected by the tracking system may be provided to one or more controllers configured to process the information and calculate a viewpoint for the firearm and/or the firearm scope.
  • image data to be presented via the display may be determined.
  • a position of the firearm may be adjusted or corrected prior to determining the image data to be presented.
  • an external display e.g., one or more projection screens, etc.
  • the position of the firearm Prior to determining a viewpoint of the firearm, the position of the firearm may be corrected to the calibration point.
  • the image data presented via the display may correspond to that presented to the user via the external display, regardless of the user's movement within a simulation environment. This position correction may assist in facilitating a simulation environment in which multiple firearms are utilized.
  • a trigger mechanism or trigger actuation detection mechanism may also be associated with the firearm.
  • a trigger mechanism that facilitates the detection of a trigger pull event may be incorporated into a firearm.
  • a trigger detection mechanism such as a false or inert round of ammunition that detects actuation of a tiring pin, may be inserted into a firearm. Based upon the detection of a trigger actuation, a firing event may be identified, and an indication of the firing event may be communicated to one or more controllers.
  • the controllers may utilize a wide variety of information associated with the firearm, scope, simulated ammunition, and/or simulated conditions (e.g., windage, elevation, etc.), as well as one or more ballistic models, to determine a ballistic path for the shot within the virtual environment. As a result of calculating a shot trajectory and/or a ballistic path, the accuracy of the simulated shot may be enhanced. Following the determination of information associated with a simulated shot, the image data output for presentation via the scope display may be updated.
  • simulated conditions e.g., windage, elevation, etc.
  • the system 100 may include, for example, a firearm 105 , a scope 110 mounted onto the firearm 105 , a display 115 associated with the scope 110 , one or more motion and/or position tracking systems 120 , and/or one or more controllers 125 .
  • the motion tracking systems 120 may be configured to track or monitor motion, position, and/or orientation information associated with the firearm. Information collected by the motion tracking systems 120 may be utilized to determine a viewpoint associated with the firearm 105 , and the determined viewpoint may be utilized to determine and/or generate image data (e.g., video data) to be rendered via the display 115 .
  • the firearm 105 may include any suitable firearm, modified firearm, or replica firearm suitable for use in a shooting or training simulation.
  • the firearm 105 may be a replica firearm (e.g., a replica or model rifle, pistol, shotgun, etc.) that has been modified to include one or more simulation components (e.g., a scope-mounted display, a scope-integrated display, one or more motion tracking systems, etc.).
  • the firearm 105 may be an actual firearm (e.g., rifle, pistol, shotgun, etc.) that has been modified to include or be otherwise associated with one or more simulation components.
  • a user may experience authentic trigger pressure, weapon weight, and/or weapon adjustments within a simulation environment.
  • components may be incorporated into the actual firearm 105 and/or removably attached to the firearm 105 .
  • simulation components and/or equipment may be integrated into a firearm to form a simulation firearm.
  • a portion of the simulation components may be integrated into a firearm while other simulation components are attached to the firearm.
  • simulation components and/or equipment may be attached to the outside of an actual weapon.
  • a user may attach simulation components to a personal firearm in order to utilize the personal firearm in a simulation environment. Indeed, a wide variety of firearm configurations may be utilized in accordance with various embodiments of the invention.
  • a plurality of firearms may be utilized in association with a training system 100 .
  • Each of the firearms may include or be associated with respective simulation components (e.g., a display, one or more motion tracking systems, etc.).
  • the system 100 includes one or more display devices situated external to the firearms, such as the one or more projection screens 130 illustrated in FIG. 1
  • the respective users associated with the firearms may be presented the same simulation or scenario by the external display devices.
  • the individual displays associated with the respective firearm scopes may present the users with individual image data.
  • the scope 110 may be any suitable scope device, modified scope device, or replica scope device suitable for use in a shooting or training simulation.
  • the scope 110 may be a replica scope that has been modified to include or otherwise be associated with a display 115 and/or other simulation components.
  • the scope 11 . 0 may be an actual scope (i.e., a telescopic sight) that has been modified to include or otherwise be associated with a display and/or other simulation components.
  • a display 115 may be associated with the scope 110 .
  • the display 115 may be at least partially mounted within a housing of the scope 110 .
  • a display 115 may be integrated into a scope 110 utilized in association with a modified simulation firearm.
  • the display 115 may be removably attached to the scope.
  • a display 11 . 5 may be at least partially mounted within the housing of a scope 110 that is attached to an actual firearm when the firearm is utilized in conjunction with the simulation system 100 .
  • the display 115 may be any suitable display device configured to present image data, such as video data, to a user of the firearm 105 .
  • suitable display devices include, but are not limited to, microdisplays, liquid crystal displays (“LCD's”), light emitting diode (“LED”) displays, organic light emitting diode (“OLED”) displays, active matrix OLED (“AMOLED”) displays, pico front or rear projection displays, stereoscopic displays, holographic displays, and/or cathode ray tube (“CRT”) displays.
  • the display 115 may be configured to present high-definition imagery.
  • the display 115 may be at least partially mounted within the housing of the scope 110 .
  • the display 115 may be removably attached to the scope 110 .
  • the display may be attached to either end of the scope 110 .
  • a video receiver 135 may be provided to route image and video signals to the display 115 for presentation to a user.
  • a video receiver 135 may receive image data from the one or more controllers 125 , and the video receiver 135 may process and/or format the received image data. The processed and/or formatted image data may then be provided to the display 115 .
  • the video receiver 135 may be either a wired or wireless video receiver.
  • the video receiver 135 may be connected to the one or more controllers 125 via a video graphics array (“VGA”) cable, an Ethernet cable, a Universal Serial Bus (“USB”) cable, a mini-USB cable, a Firewire cable, a digital video interface (“DVI”) cable, a high definition multimedia interface (“HDMI”) cable, or other suitable cable or wired connection.
  • VGA video graphics array
  • USB Universal Serial Bus
  • DVI digital video interface
  • HDMI high definition multimedia interface
  • the video receiver 135 may be connected to the one or more controllers 125 via a suitable Bluetooth connection, wireless Ethernet connection, Wi-Fi connection, or other wireless connection.
  • one or more suitable communications interfaces 140 may facilitate communications between the video receiver 135 and the one or more controllers 125 .
  • the video receiver 135 may be incorporated and/or integrated into the firearm 105 or the scope 110 . In other embodiments, the video receiver 135 may be removably attached to the firearm 105 .
  • an audio video (“AV”) receiver may be provided, and the AV receiver may process both audio and video data.
  • the video data may be provided to the display 110 , and the audio data may be provided to one or more suitable audio presentation devices (e.g., speakers, headphones, etc.).
  • the AV receiver may receive separate audio and video signals.
  • separate audio and video signals may be received from the one or more controllers 125 .
  • a video signal may be received from the one or more controllers 125
  • a separate audio signal may be received from one or more other controllers associated with the generation and presentation of a primary or main simulation scenario, such as a scenario presented via the one or more projection screens 130 .
  • a motion tracking system 120 may be configured to monitor and/or track the motion, position, and/or orientation of the firearm 105 and/or the scope 110 .
  • a viewpoint or perspective of the firearm 105 and/or the scope 110 may be determined, and image data may be determined and/or generated for presentation by the display 115 .
  • suitable motion tracking systems 120 include, but are not limited to, a laser tracking system, an electromagnetic tracking system, an optical tracking system, an infrared tracking system, an ultrasonic tracking system, an inertia tracking system, a mechanical tracking system, and/or a Global Positioning System (“GPS”) tracking system.
  • GPS Global Positioning System
  • a single motion tracking system may be utilized. In other embodiments, a plurality or combination of motion tracking systems may be utilized. In this regard, positional and/or orientation information for the firearm 105 and/or the scope 110 may be verified and/or accuracy may be improved.
  • a laser tracking system may include a laser emitter that is attached to or incorporated into the firearm 105 or the scope 110 .
  • a laser emitter may be inserted into the barrel of the firearm 105 .
  • One or more laser detection devices such as a laser camera or other laser detector, may also be attached to the firearm 105 and/or at one or more other positions suitable for viewing an area in which the laser emitter is pointed.
  • one or more laser detection devices may be utilized to identify an emitted laser that contacts and/or that is reflected off of the one or more projection screens 130 . Based upon the capture of the emitted laser, positional and/or orientation data for the firearm 105 may be determined or calculated.
  • An electromagnetic tracking system may include one or more electromagnetic sources attached to the firearm 105 and one or more sensors utilized to monitor a calibrated area for position and motion of the electromagnetic sources.
  • positional and/or orientation data for the firearm 105 may be determined or calculated.
  • a video device may be calibrated to one or more unique optical markers attached to the firearm 105 and/or to one or more visual identifiers inherent to the firearm 105 .
  • Motion of the firearm 105 may be monitored by the video device, and information associated with the monitored motion may be translated into positional and/or orientation data for the firearm 105 .
  • a series of LEDs and/or reflective markers may be attached to the firearm 105 .
  • the LEDs may be utilized to emit infrared signals that are reflected by the markers.
  • One or more infrared cameras may be utilized to monitor a field of view encompassing the firearm, such as a simulation area. As the firearm 105 is moved, the infrared camera may identify movement of the LEDs and/or reflective markers, and information associated with the identified movement may be translated into positional and/or orientation data for the firearm 105 . In certain embodiments, movement of the LEDs and/or reflective markers may be identified relative to one another in order to provide a perspective for the firearm 105 .
  • An ultrasonic tracking system may include one or more ultrasonic transmitters that are attached to the firearm 105 .
  • One or more audio sensors situated in predetermined positions may be utilized to measure the time-of-flight and/or the phase of the ultrasonic sound emitted by the ultrasonic transmitters. The measurements data may then be processed in order to determine distances between the sensors and the firearm, and the distances may be utilized to determine positional and/or orientation data for the firearm 105 .
  • one or more inertia sensors including gyro meters and/or accelerometers may be attached to the firearm 105 . Motion or movement of the firearm 105 may be measured as inertia and translated into positional and/or orientation data for the firearm 105 .
  • a mechanical tracking system may include one or more mechanical devices and/or mechanical sensors that are attached to the firearm 105 such that movement of the firearm 105 directly affects the mechanical devices. Movement of the gun may be translated into movement of portions of the mechanical devices and may be translated into positional and/or orientation data.
  • a GPS tracking system one or more GPS receivers may be attached to the firearm 105 .
  • a GPS receiver may precisely measure the timing of signals output by orbital GPS satellites. As the firearm 105 moves, the GAS receiver continually calculates positional information in real-time and tracks movement of the firearm 105 . The movement data may then be translated into positional and/or orientation information for the firearm 105 .
  • movement and/or motion of the firearm 105 may be tracked along any number of axes utilizing two to six degrees of freedom.
  • motion may be tracked along an x, y, and/or z axis.
  • two, three, four, five, or six degrees of freedom may be utilized to track or monitor motion of the firearm 105 .
  • motion may be selectively tracked as the firearm is moved forward and backward (e.g., closer to or further away from a projection screen), left or right, and/or up or down.
  • yaw, pitch, and/or roll of the firearm 105 may be tracked.
  • the tracking of the firearm may be utilized to determine a viewpoint and/or perspective of the firearm in order to determine image data for presentation via the display 115 .
  • one or more communications interfaces 140 may be incorporated into and/or attached to the firearm 105 or the scope 110 .
  • the communications interfaces may include wired and/or wireless communications devices (e.g., transceivers, communications cards, etc.) that facilitate communication with the controllers 125 .
  • data and/or information collected by simulation components attached to the firearm 105 e.g., the motion tracking systems 120 , a camera, a trigger recognition device, a keypad, etc.
  • data output by the controllers 125 e.g., image data, calibration data etc.
  • a camera 145 may be integrated into or incorporated into the firearm 105 or the scope 110 .
  • the camera 145 may be utilized to detect and/or monitor grid patterns projected onto or output by an external display surface, such as the one or more projection screens 130 .
  • the data captured by the camera 145 may be processed in order to determine an orientation and/or perspective of the firearm 105 . In this regard, the motion tracking of the firearm 105 may be verified. Additionally, a determination may be made as to whether the firearm 105 is positioned correctly.
  • one or more trigger mechanisms 150 and/or trigger detection devices may be incorporated into, inserted into, and/or attached to the firearm 105 . These mechanisms and/or devices may facilitate the detection of a trigger pull or a trigger actuation. Once a trigger actuation has been detected, an indication of the detected actuation may be communicated to the controllers 125 , and a shot event may be registered.
  • a wide variety of trigger mechanisms 150 and/or trigger detection devices may be utilized as desired in various embodiments of the invention.
  • a firearm 105 may be modified to include a trigger mechanism that includes one or more actuation sensors.
  • an inert or dummy round may be inserted into a firearm 105 , and the dummy round may include one or more suitable electrical components that detect actuation of the firing pin.
  • An inert round is explained in greater detail below with reference to FIGS. 2A and 2B .
  • the trigger mechanisms and/or trigger detection devices described above are provided by way of example only, and other suitable devices may be utilized as desired in various embodiments of the invention.
  • sensors may be incorporated into and/or attached to the firearm 105 as desired in various embodiments.
  • one or more sensors may be attached to the scope 110 in order to detect user adjustment made to the scope, such as windage adjustments, elevation adjustments, focus adjustments, zoom adjustments, and/or mode adjustments (e.g., a standard mode, a thermal imaging mode, a night vision mode, etc.).
  • one or more sensors may be utilized to monitor the removal and/or insertion of magazines.
  • information collected by the sensors may be communicated to the controllers 125 and utilized during the presentation of audio and/or image data during a simulation.
  • any number of controllers 125 may be provided.
  • a controller 125 may be situated or positioned external to the firearm 105 and/or the scope 110 .
  • a controller 125 may be incorporated into and/or attached to the firearm 105 and/or the scope 110 .
  • Each controller 125 may be a suitable processor-driven device that facilitates the provision of a firearms simulation.
  • a controller 125 may be a suitable processor-driven device that receives motion information associated with the firearm 105 , determines a viewpoint of the firearm 105 , generates or determines image data for presentation via the display 115 , and/or outputs the image data for presentation via the display 115 .
  • the controller 125 may include any number of computing devices, such as a personal computer, a digital assistant, a personal digital assistant, a digital tablet, an Internet appliance, an application-specific circuit, a microcontroller, a minicomputer, or any other processor-based device.
  • the execution of suitable computer-implemented instructions or computer-executable instructions by the controller 125 may form a special purpose computer or other particular machine that is operable to facilitate the provision of a firearms simulation.
  • a controller 125 may include one or more processors 152 , one or more memory devices 154 , one or more network or communications interfaces 156 , and/or one or more input/output (“I/O”) interfaces 158 .
  • the processors 152 may be configured to execute any number of software applications and/or computer-readable or computer-executable instructions.
  • the memory devices 154 may include any number of suitable memory devices, such as caches, read-only memory devices, random access memory devices, flash memory devices, magnetic storage devices, removable storage devices (e.g., memory cards, memory sticks, etc.), etc.
  • the memory devices 154 may include internal memory devices and/or external memory devices in communication with the controller 1 . 25 .
  • the memory devices 154 may store data, executable instructions, and/or various program modules utilized by the processors 152 , such as data files 160 , an operating system (“OS”) 162 , a tracking module 164 , a video module 166 , and/or a ballistics module 168 .
  • OS operating system
  • the memory devices 154 may store data, executable instructions, and/or various program modules utilized by the processors 152 , such as data files 160 , an operating system (“OS”) 162 , a tracking module 164 , a video module 166 , and/or a ballistics module 168 .
  • OS operating system
  • the data files 160 may include any suitable data that facilitates the operation of the controller 125 and/or the provision of a firearms simulation.
  • the data files 160 may include, but are not limited to, received motion and/or movement data for one or more firearms 105 , calibration information for a simulation environment, information associated with types of rounds or ammunition to be simulated, information associated with desired scope settings, information associated with desired simulation conditions (e.g., wind, etc.), one or more ballistic models, stored image and/or audio data, and/or information utilized to generate and/or format image and/or audio data for presentation during a simulation.
  • the OS 162 may be a suitable software module that controls the general operation of the controller 125 and/or the execution of other software modules, for example, the tracking module 164 , the video module 166 , and/or the ballistics module 168 .
  • the tracking module 164 may be a suitable software module or application that facilitates the determination of a viewpoint of the firearm 105 .
  • the viewpoint of the firearm 105 may then be utilized to determine video to be presented via the display 115 .
  • the viewpoint of the firearm 105 may take a wide variety of position, orientation, and/or perspective information into account, as well as various settings associated with the scope 110 (e.g., a magnification setting, an operational mode, etc.).
  • the tracking module 164 may receive measurements and/or tracking data from the one or more motion tracking systems 120 , and the tracking module 164 may utilize at least a portion of the received data to determine a viewpoint of the firearm 105 .
  • received measurements data may be translated into positional and/or orientation information for the firearm 105 .
  • the positional and/or orientation information may then be utilized to determine a viewpoint for the firearm 105 .
  • an area of the one or more projection screens 130 at which the firearm 105 is aimed or pointed may be determined.
  • various tilts and/or angles of the firearm e.g., pitch, yaw, roll
  • a distance between the firearm 105 and a relevant projection screen 130 may be taken into consideration when determining the viewpoint.
  • the tracking module 164 may receive tracking data from a plurality of motion tracking systems 120 , and the tracking module 164 may process the respective data to verify the positional and/or orientation information for the firearm 105 .
  • positional and/or orientation information may be averaged or otherwise processed in the event that tracking data is received from a plurality of sources.
  • grid pattern information may be received from the camera 145 and utilized to verify a viewpoint of the firearm 105 .
  • the tracking module 164 may correct or adjust respective positions of the one or more firearms.
  • image data that is determined or generated for display via the firearm scopes may correspond to the perspective illustrated via the external projection screens 130 , which is the same for all users within the simulation environment.
  • the position of the firearm 105 may be corrected to a calibration point for the simulation environment.
  • the image data presented via the display 115 may correspond to that presented to the user via the projection screens 130 , regardless of the user's movement within a simulation environment.
  • a simulation scenario illustrated on the one or more external displays may be modified as the user moves within a simulation environment.
  • a personal external display e.g., a personal monitor, a headset, etc.
  • the video module 166 may be a suitable software module or application configured to determine and/or generate image data for presentation via the display 115 .
  • the video module 166 may receive viewpoint information for the firearm 105 , and the video module 166 may utilize the viewpoint information to generate image data for communication to the video receiver 135 and/or the display 115 .
  • the video module 166 may additionally utilize one or more settings associated with the scope (e.g., a magnification, a display mode, etc.) and/or an indication of a shot event to determine or generate image data for presentation.
  • a wide variety of image data may be generated by the video module 166 , including simulation scenario imagery, enhanced or enlarged imagery corresponding to a portion or section of the imagery displayed on the projection screens 130 , imagery associated with a simulated view through the scope 110 (e.g., a reticle, etc.), and/or imagery associated with a simulated shot.
  • the video module 166 may also include an audio component configured to determine or generate audio for presentation to a user of the firearm 105 .
  • the ballistics module 168 may be a suitable software module or application configured to calculate or determine a simulated trajectory of a shot within a virtual environment. Once a shot event has been identified, information associated with the viewpoint and/or orientation of the firearm 105 may be provided to the ballistics module 168 , and the ballistics module 168 may calculate a trajectory for the shot. As desired, the ballistics module 168 may utilize a wide variety of other information to calculate a trajectory for the shot. For example, information associated with the firearm 105 (e.g., caliber information, etc.), information associated with a simulated ammunition, elevation information and/or wind information may be provided to any number of suitable ballistic models to calculate or determine a trajectory for the simulated shot.
  • Results of the simulated shot may then be provided to the video module 166 to facilitate the generation of image data associated with the simulated shot.
  • shot accuracy may be improved or enhanced.
  • conventional training systems such as systems that detect a shot by detecting a laser projected onto a two-dimensional surface
  • the size of the laser point or other shot may be bigger than the target aimed at, thereby reducing accuracy and ballistic fidelity.
  • external shot detection mechanisms may be bypassed, and overall accuracy may be improved.
  • a wide variety of suitable operations may be performed by the simulation components associated with a firearm 105 and/or by the one or more controllers 125 to facilitate provision of a training simulation.
  • the various operations described above for a controller 125 are provided by way of example only. Another example of the operations that may be performed by the controllers 125 is described in greater detail below with reference to FIG. 8 . Additionally, an example of the operations that may be performed by the simulation components associated with the firearm 105 is described in greater detail below with reference to FIG. 7 .
  • the one or more I/O interfaces 158 may facilitate communication between the controller 125 and one or more input/output devices, for example, one or more user interface devices, such as a display, keypad, mouse, pointing device, control panel, touch screen display, remote control, microphone, speaker, etc., that facilitate user interaction with the controller 125 .
  • user interface devices such as a display, keypad, mouse, pointing device, control panel, touch screen display, remote control, microphone, speaker, etc.
  • user commands may be received by the controller 125 , and a wide variety of data may be output for presentation to a user.
  • the one or more network interfaces 156 may facilitate connection of the controller 125 to one or more suitable wired and/or wireless networks 145 , for example, a local area network, a Wi-Fi enabled network, a Bluetooth-enabled network, a radio frequency network, the Internet, etc. Communications between various components of the system 100 may be facilitated via the networks 145 .
  • one or more projection screens 130 and/or other external display devices may be utilized in conjunction with the training system 100 .
  • one or more projection screens 130 may be utilized to present a simulation scenario (e.g., a military training exercise, a law enforcement training exercise, a hunting scenario, etc.) to any number of users associated with respective firearms.
  • a simulation scenario e.g., a military training exercise, a law enforcement training exercise, a hunting scenario, etc.
  • a wide variety of projection screen arrangements may be utilized in various embodiments of the invention.
  • a plurality of screens may be placed next to each other to enlarge a simulation viewing area.
  • the projections screens may positioned in a curve to create up to a 360 degree simulation viewing area. Indeed, projection screens may even be positioned in a half dome or full dome configuration.
  • one or more projectors 170 or other image generation devices may be configured to output images and/or video presented via the projection screens 130 .
  • a simulation controller 175 may be a suitable processor-driven device configured to control the presentation of a simulation via the one or more projection screens 130 .
  • a simulation controller 175 may include components similar to those described above for the controller 125 .
  • a simulation controller 175 may include one or more processors and/or memory devices, and simulation software may be executed by the processors to facilitate the generation of image data to be. displayed via the projection screens 130 .
  • a simulation controller 175 may be separate from one or more controllers 125 associated with individual firearms. In other embodiments, the simulation controller 175 may be combined with one or more other controllers 125 .
  • system 100 shown in and described with respect to FIG. 1 is provided by way of example only. Numerous other operating environments, system architectures, and device configurations are possible. Other system embodiments can include fewer or greater numbers of components and may incorporate some or all of the functionality described with respect to the system components shown in FIG. 1 .
  • FIGS. 2-5 illustrate a few example modified firearms that may be utilized.
  • FIG. 2A is an illustration of an actual firearm 200 that has been modified to include attached simulation components.
  • the firearm 200 may be any suitable firearm onto which a scope 205 has been mounted.
  • the scope 205 may be an actual scope mounted to the firearm 200 .
  • a scope attachment 210 may be connected (e.g., removably connected) to one end of the scope 205 to facilitate the presentation of image data to a user.
  • the scope attachment 210 may include a suitable visual display unit configured to present image data.
  • a visual display unit may include, for example, a lens, a digital display micro-screen or other suitable display, a control board, and/or a communications interface.
  • FIG. 2B is an exploded illustration of the firearm 200 shown in FIG. 2A . Also shown in FIG. 2B is the interior 225 of the inert round 220 . The interior 225 may be a spring-mounted metal bolt.
  • a circuit 230 such as an approximately five (5) volt circuit, may be connected to the inert round 220 and grounded to the firearm 200 . A trigger pull even may lead to the inert round being contacted by the firing pin of the firearm 200 , thereby causing the circuit 230 to detect a shot being fired.
  • FIGS. 2A-2B illustrate an actual firearm, replica or model firearms may be modified in a similar manner for use in a simulation system.
  • FIG. 3 illustrates a block diagram of an example scope 300 that may be mounted on a firearm in accordance with various embodiments of the invention.
  • the scope 300 may include any number of incorporated, integrated, and/or attached simulation components.
  • the scope 300 may include a suitable display device 305 , a video receiver 310 , and/or one or more communications interfaces 315 .
  • Each of these components may be similar to the components described in greater detail above with reference to FIG. 1 .
  • These components may facilitate the receipt of image data output by a controller, and the presentation of the image data to a user.
  • the communications interfaces 315 associated with the scope 300 may also be utilized to facilitate communication of measurements data to one or more controllers.
  • the communications interfaces 315 may be in communication with a trigger mechanism 320 , a motion tracking system 325 , and/or a camera 330 .
  • shot indication data, motion data, and/or grid pattern data may be communicated to one or more controllers for processing.
  • FIG. 3 illustrates a motion tracking system 325 and camera 330 that have been integrated into the scope 300 , one or more of these components may be external to the scope 300 .
  • FIG. 4 is an illustration of an actual firearm 400 onto which a replica scope 405 has been mounted to facilitate the use of the firearm 400 within a simulation environment.
  • the replica scope 405 may include components similar to those described above with reference to FIG. 3 .
  • one or more tracking sensors 410 may be attached to the firearm 400 to facilitate motion tracking.
  • an inert round 415 may be inserted into the firearm to facilitate detection of shot events.
  • FIG. 5 is an illustration of a firearm 500 into which simulation components have been integrated.
  • the firearm 500 illustrated in FIG. 5 may be a replica firearm or an actual firearm that has been modified to include integrated simulation components.
  • the firearm 500 may be modified to include a trigger mechanism, a tracking system, and/or a communications interface.
  • the scope may be modified to include a display device and a video receiver.
  • FIG. 6 is a diagram of a simulation environment 600 that may be utilized in accordance with certain embodiments of the invention.
  • the simulation environment 600 may include any number of displays 605 , such as projection screens, that facilitate presentation of simulation scenarios. Although three projection screens are illustrated in FIG. 6 , any number of suitable displays may be utilized. For example, a relatively circular simulation environment may be provided in which displays surround users in three hundred and sixty degrees. Additionally, as desired, any number of firearms 610 a - n may be utilized in conjunction with the simulation environment 600 . Each of the firearms 610 a - n may be associated with a respective scope-integrated or scope-mounted display. Thus, users may view a simulation scenario within the simulation environment 600 via the external displays 605 . Each user may then use a scope-mounted display associated with his/her firearm to view image data determined based upon the viewpoint of the firearm.
  • a perspective may be predetermined or assumed for the external displays 605 .
  • a calibration point 615 may be situated at a relatively central point for viewing the external displays 605 .
  • the firearms may be calibrated for use with the simulation system 600 at the calibration point 615 .
  • a user may move a firearm, such as firearm 610 a, within the simulation environment 600 .
  • the firearm 610 a may be pointed at different portions of the external displays 605 . Accordingly, the viewpoint of the firearm 610 a may change as the firearm 610 a is moved.
  • the position of the firearm 610 a within the simulation environment 600 may be autocorrected to the calibration point 615 . For example, an area or portion of the external displays 605 at which the firearm 610 a is aimed may be determined.
  • a viewpoint of that area or portion from the perspective of the calibration point may then be determined as the viewpoint of the firearm 610 a.
  • the imagery illustrated via the scope-mounted display may correlate or correspond to the imagery presented to the user via the external displays 605 .
  • FIG. 7 is a flow diagram of an example method 700 for presenting image data via a display associated with a scope mounted onto a firearm, according to an example embodiment of the invention.
  • the method 700 may be performed by a suitable firearm, such as the firearm 105 illustrated in FIG. 1 .
  • the method 700 may begin at block 705 .
  • communication may be initiated with a host controller, such as a host computer.
  • a host controller such as a host computer.
  • motion tracking information may be communicated to the host controller, and the image data may be received from the host controller for presentation via the display associated with the scope.
  • a tracking system may be configured to track motion and/or movement of the firearm 105 within a simulation environment.
  • a tracking system may be configured to determine position and/or orientation information for the firearm 105 .
  • position and/or orientation information for the firearm 105 may be determined, and the position and/or orientation information may be communicated to the host controller at block 720 .
  • tracking information collected by a tracking system may be provided or communicated to the host controller at block 720 , and the host controller may determine position and/or orientation information for the firearm 105 .
  • a viewpoint of the firearm 105 or the firearm scope may be determined by the host controller, and the host controller may generate image data in accordance with the determined viewpoint.
  • the image data may be output by the host controller for receipt by the firearm 105 , and the firearm 105 may receive the image data at block 725 .
  • the received image data such as a received image feed, may then be output by the display for presentation to the user at block 730 . Accordingly, when a user views the scope associated with the firearm 105 , the user may view the image data output by the host controller.
  • a determination may be made as to whether the image is calibrated. In other words, a determination may be made as to whether the firearm 105 has been properly calibrated within the simulation system and/or the virtual environment. If it is determined at block 735 that the image is not calibrated, then operations may continue at block 740 .
  • the firearm 105 and/or the position of the firearm 105 may be calibrated within the virtual environment. For example, a user may be prompted to utilize a suitable calibration routine (e.g., aiming the firearm 105 at the four corners of an external display, etc.) to calibrate the firearm 105 . Operations may then continue at block 725 described above.
  • operations may continue at block 745 .
  • a determination may be made as to whether a trigger associated with the firearm 105 has been pulled or actuated by a user. In other words, a determination may be made as to whether a shot has occurred.
  • suitable trigger mechanisms and/or trigger detection devices e.g., an inert round, etc.
  • operations may continue at block 765 described in greater detail below. If, however, it is determined at block 745 that the trigger has been actuated or pulled, then operations may continue at block 750 .
  • a shot indication may be communicated from the firearm 105 to the host controller.
  • the shot indication may be processed by the host controller to simulate a shot within a virtual environment.
  • one or more ballistic models may be utilized to determine a trajectory of a shot within the virtual environment.
  • Image data e.g., an image feed
  • the firearm 105 may receive the updated image data at block 755 and present information associated with a virtual shot via the display at block 760 . Operations may then continue at block 765 .
  • a determination may be made as to whether a position and/or orientation of the firearm 105 has changed. For example, a determination may be made as to whether the firearm 105 has been moved or whether a tracking system has detected movement of the firearm 105 . If it is determined at block 765 that the position and/or orientation of the firearm 105 has not changed, then operations may continue at block 730 , and image data may continue to be output. If, however, it is determined at block 765 that the position and/or orientation of the firearm 105 has changed, then operations may continue at block 715 (or block 720 ). In this regard, a new viewpoint of the firearm 105 may be determined, and the image data output by the display may be updated to reflect the new viewpoint.
  • the method 700 may end following the completion of a simulation scenario in which the firearm 105 is utilized.
  • FIG. 8 is a flow diagram of an example method 800 for determining image data to be presented via a display associated with a scope mounted onto a firearm, according to an example embodiment of the invention.
  • the method 800 may be performed by one or more suitable controllers associated with a simulation system, such as the controllers 125 illustrated in FIG. 1 .
  • the method 800 may begin at block 805 .
  • the controller 125 may initiate communication with a firearm and/or a scope, such as the firearm 105 and/or the scope 110 illustrated in FIG. 1 .
  • the controller 125 may receive a wide variety of data and/or indications from the firearm 105 , such as indications associated with shot events or trigger actuations and/or information associated with scope settings.
  • a communications channel may also be established for communicating image data to the firearm 105 and/or the scope 110 .
  • one or more firearm and/or scope parameters may be determined.
  • the one or more parameters may be determined or identified within suitable ballistics simulation software, such as the ballistics module 168 illustrated in FIG. 1 .
  • suitable ballistics simulation software such as the ballistics module 168 illustrated in FIG. 1 .
  • a wide variety of different firearm and/or scope parameters may be determined as desired in various embodiments of the invention, such as a type associated with the firearm 105 (e.g., a caliber, etc.), a ballistic profile associated with the firearm 105 , a type of ammunition to be simulated, a ballistic profile associated with the ammunition, a zoom or magnification for the scope 110 , a mode of operation for the scope 110 (e.g., normal operation, an infrared mode, a night vision mode, etc.), and/or windage and/or elevation adjustments for the scope 110 , etc.
  • the trajectory of simulated shots may be determined in a relatively accurate manner.
  • the controller 125 may initiate communication with one or more suitable tracking systems associated with the firearm 105 and/or the scope 110 , such as the motion tracking system 120 illustrated in FIG. 1 .
  • the controller 125 may receive motion, position, and/or orientation data associated with the firearm 105 .
  • position, orientation, and/or viewpoint information for the firearm 105 may be determined. For example, tracking data received from the motion tracking systems 120 may be translated into position and/or orientation data for the firearm 105 . A viewpoint of the firearm 105 within the virtual environment may then be determined.
  • a position of the firearm 105 within the virtual environment may be corrected or adjusted.
  • the position of the firearm 105 may be set to a predetermined position, such as a predetermined calibration point.
  • a determined viewpoint for the firearm 105 may correspond with the viewpoint presented via one or more external (i.e., non-firearm-mounted) displays.
  • an image feed for display via a scope-mounted display may be determined and/or generated based at least in part upon the determined viewpoint for the firearm 105 .
  • the image feed may then be output at block 835 for receipt by the firearm 105 and presentation to the user via the scope-mounted display.
  • a determination may be made as to whether the image communicated to the firearm 105 is calibrated. In other words, a determination may be made as to whether the firearm 105 has been properly calibrated within the simulation system and/or the virtual environment. If it is determined at block 840 that the image is not calibrated, then operations may continue at block 845 .
  • the firearm 105 and/or the position of the firearm 105 may be calibrated within the virtual environment. For example, a user may be prompted to utilize a suitable calibration routine (e.g., aiming the firearm 105 at the four corners of an external display, etc.) to calibrate the firearm 105 . Operations may then continue at block 830 described above.
  • operations may continue at block 850 .
  • a determination may be made as to whether a shot event has occurred. For example, a determination may be made as to whether a shot indication has been received from the firearm 105 . If it is determined at block 850 that a shot event has not occurred, then operations may continue at block 865 described in greater detail below. lf, however, it is determined at block 850 that a shot event has occurred, then operations may continue at block 855 .
  • the shot event may be registered within the ballistic simulation software, and a trajectory of the shot may be calculated or determined.
  • a trajectory of the shot may be calculated or determined.
  • ballistic factors such as weather conditions, elevation, ammunition type, firearm type, etc.
  • operations may continue at block 860 , and the image feed may be adjusted or modified to correspond with the registered shot.
  • Image information associated with the simulated shot may then be output for receipt via the firearm 105 and presentation via the scope-mounted display. Operations may then continue at block 865 .
  • a determination may be made as to whether a position and/or orientation of the firearm 105 has changed. For example, a determination may be made as to whether motion or movement information has been received from one or more tracking systems. As another example, a determination may be made as to whether updated position and/or orientation information has been received. If it is determined at block 865 that the position and/or orientation of the firearm 105 has not changed, then operations may continue at block 835 , and image data may continue to be output for receipt by the firearm 105 . If, however, it is determined at block 865 that the position and/or orientation of the firearm 105 has changed, then operations may continue at block 820 , and a new viewpoint for the firearm 105 may be determined. In this regard, the image data output for display by the scope may be updated to reflect the new viewpoint.
  • the method 800 may end following the completion of a simulation scenario associated with the controller 125 .
  • the operations described and shown in the methods 700 and 800 of FIGS. 7-8 may be carried out or performed in any suitable order as desired in various embodiments of the invention. Additionally, in certain embodiments, at least a portion of the operations may be carried out in parallel. Furthermore, in certain embodiments, less than or more than the operations described in FIGS. 7-8 may be performed.
  • These computer-executable program instructions may be loaded onto a special purpose computer or other particular machine, a processor, or other programmable data processing apparatus to produce a particular machine, such that the instructions that execute on the computer, processor, or other programmable data processing apparatus create means for implementing one or more functions specified in the flow diagram block or blocks.
  • These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means that implement one or more functions specified in the flow diagram block or blocks.
  • embodiments of the invention may provide for a computer program product, comprising a computer-usable medium having a computer-readable program code or program instructions embodied therein, said computer-readable program code adapted to be executed to implement one or more functions specified in the flow diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational elements or steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process such that the instructions that execute on the computer or other programmable apparatus provide elements or steps for implementing the functions specified in the flow diagram block or blocks.
  • blocks of the block diagrams and flow diagrams support combinations of means for performing the specified functions, combinations of elements or steps for performing the specified functions and program instruction means for performing the specified functions. It will also be understood that each block of the block diagrams and flow diagrams, and combinations of blocks in the block diagrams and flow diagrams, can be implemented by special purpose, hardware-based computer systems that perform the specified functions, elements or steps, or combinations of special purpose hardware and computer instructions.

Abstract

Firearm training systems and associated methods are provided. A firearm training system may include a firearm, a scope, a tracking system, and at least one controller. The scope may be mounted to the firearm, and the scope may include a display. The tracking system may be configured to collect information associated with a position and an orientation of the firearm. The at least one controller may be configured to (i) determine, based at least in part upon the information collected by the tracking component, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • The present application claims the benefit of U.S. Provisional Application Ser. No. 61/308,169, filed Feb. 25, 2010, and entitled “Firearm Scope and System for Simulated Environments,” the contents of which are incorporated by reference herein in their entirety.
  • FIELD OF THE INVENTION
  • Aspects of the invention relate generally to firearm training systems and, more particularly, to firearm training systems that utilize a simulated virtual environment.
  • BACKGROUND OF THE INVENTION
  • Training for marksmen and snipers, including military and law enforcement personnel, is often critical for the development of targeting and accuracy skills. However, live training often requires relatively expensive ammunition and the maintenance of shooting range facilities. Additionally, live training typically fails to provide a variety of targets, including moving targets. In order to provide additional avenues for firearms training, various shooting simulators have been developed. These conventional shooting simulators typically present images via projection screens or other image displays. Simulated projectiles or lasers are then typically identified as shots when contact is made with the display surface.
  • However, due to the utilization of two-dimensional surfaces for shot recognition and capture, accuracy is typically compromised in these conventional shooting simulators. These conventional shooting simulators also fail to accurately represent ballistic physics during the capture of shooting events. Additionally, these conventional shooting simulators fail to effectively utilize scopes to represent digital imagery within a virtual environment. For example, if a user were to look through a scope at an image display utilized in a conventional shooting simulator, the user will be presented with a pixilated magnified view of the image display. Accordingly, improved firearm training systems and associated methods are desirable.
  • BRIEF DESCRIPTION OF THE INVENTION
  • Some or all of the above needs and/or problems may be addressed by certain embodiments of the invention. Embodiments of the invention may include firearm training systems and methods of using the same. In one embodiment, a firearm training system may be provided. The firearm training system may include a firearm, a scope, a tracking system, and at least one controller. The scope may be mounted to the firearm, and the scope may include a display. The tracking system may be configured to collect information associated with a position and an orientation of the firearm. The at least one controller may be configured to (i) determine, based at least in part upon the information collected by the tracking component, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.
  • In accordance with another embodiment of the invention, a system for modifying a firearm for use in a simulation environment may be provided. The system may include a display, a tracking component, and at least one controller. The display may be configured to be attached to a scope associated with the firearm. The tracking component may be configured to be attached to the firearm, and the tracking component may be further configured to collect information associated with a position and an orientation of the firearm. The at least one controller may be configured to (i) determine, based at least in part upon the information collected by the tracking system, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.
  • In accordance with yet another embodiment of the invention, a firearm scope may be provided. The firearm scope may include a housing, a display, and a video receiver. The housing may be configured to be mounted onto a firearm. The display may be mounted at least partially within the housing. The video receiver may be in communication with the display and configured to output video to be rendered by the display. The video receiver may receive the video from at least one controller configured to (i) determine a viewpoint of the firearm, (ii) adjust a position of the firearm within a virtual environment, (iii) determine the video to be rendered, and (iv) direct communication of the video to the video receiver.
  • In accordance with yet another embodiment of the invention, a method for providing a firearm simulation may be provided. Motion of a firearm may be tracked via a tracking component attached to the firearm. Based upon information collected by the tracking component, a viewpoint of the firearm may be determined by a controller, such as a controller including one or more computers. Based at least in part upon the determined viewpoint, the controller may determine or identify image data to be rendered by a display associated with a scope mounted onto the firearm. The determined image data may be output by the controller for receipt by the display, and the image data may be presented by the display.
  • Additional systems, methods, apparatus, features, and aspects may be realized through the techniques of various embodiments of the invention. Other embodiments and aspects of the invention are described in detail herein with reference to the description and to the drawings and are considered a part of the claimed invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
  • FIG. 1 illustrates a block diagram of an example firearm training system that may be utilized in accordance with various embodiments of the invention.
  • FIG. 2A is an illustration of an actual firearm that has been modified to include attached simulation components.
  • FIG. 2B is an exploded illustration of the firearm shown in FIG. 2A.
  • FIG. 3 illustrates a block diagram of an example scope that may be mounted on a firearm in accordance with various embodiments of the invention.
  • FIG. 4 is an illustration of an actual firearm onto which a replica scope has been mounted to facilitate use of the firearm within a simulation environment.
  • FIG. 5 is an illustration of a firearm into which simulation components have been integrated.
  • FIG. 6 is a diagram of a simulation environment that may be utilized in accordance with various embodiments of the invention.
  • FIG. 7 is a flow diagram of an example method for presenting image data via a display associated with a scope mounted onto a firearm, according to an example embodiment of the invention.
  • FIG. 8 is a flow diagram of an example method for determining image data to be presented via a display associated with a scope mounted onto a firearm, according to an example embodiment of the invention.
  • DETAILED DESCRIPTION
  • Embodiments of the invention now will be described more fully hereinafter with reference to the accompanying drawings, in which embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Like numbers refer to like elements throughout.
  • Embodiments of the invention may include firearm training systems, apparatus, and methods for using the same. In certain embodiments, a scope may be mounted to a firearm, and a display may be associated with the scope. For example, the display may be integrated into, incorporated into, or at least partially situated within the scope. As another example, the display may be attached to the scope. A wide variety of suitable displays may be utilized as desired in various embodiments of the invention, such as a microdisplay, a liquid crystal display (“LCD”), a light emitting diode (“LED”) display, an organic light emitting diode (“OLED”) display, or other suitable display. The display may facilitate the presentation of image data, such as video, to a user of the firearm. For example, in a virtual environment, one or more displays external to the firearm (e.g., projection screens, etc.) may be utilized to present a virtual environment to a user. When a user looks through the scope, an enlarged representation of a portion of the virtual environment may be presented to the user via the display.
  • Additionally, at least one tracking system may be associated with the firearm. The tracking system may be configured to track motion, position, and/or orientation data associated with the firearm. A wide variety of suitable tracking systems may be utilized as desired in various embodiments of the invention, including but not limited to, a laser tracking system, an electromagnetic tracking system, a magnetic tracking system, an optical tracking system, an infrared tracking system, an ultrasonic tracking system, an inertia tracking system, a mechanical tracking system, a Global Positioning System (“GPS”) tracking system, or a combination of tracking systems. As desired in various embodiments, motion and/or movement of the firearm and/or scope may be tracked along any number of axes and/or in a wide variety of different directions. For example, movement may be tracked along three axes (e.g., the x, y, and z axes) with anywhere from two to six degrees of freedom.
  • Information collected by the tracking system may be utilized to determine a viewpoint associated with the firearm. For example, information collected by the tracking system may be provided to one or more controllers configured to process the information and calculate a viewpoint for the firearm and/or the firearm scope. In this regard, image data to be presented via the display may be determined. As desired in various embodiments of the invention, a position of the firearm may be adjusted or corrected prior to determining the image data to be presented. For example, an external display (e.g., one or more projection screens, etc.) may be configured to present a virtual environment from the perspective of a single calibration point. Prior to determining a viewpoint of the firearm, the position of the firearm may be corrected to the calibration point. In this regard, the image data presented via the display may correspond to that presented to the user via the external display, regardless of the user's movement within a simulation environment. This position correction may assist in facilitating a simulation environment in which multiple firearms are utilized.
  • A trigger mechanism or trigger actuation detection mechanism may also be associated with the firearm. For example, a trigger mechanism that facilitates the detection of a trigger pull event may be incorporated into a firearm. As another example, a trigger detection mechanism, such as a false or inert round of ammunition that detects actuation of a tiring pin, may be inserted into a firearm. Based upon the detection of a trigger actuation, a firing event may be identified, and an indication of the firing event may be communicated to one or more controllers. The controllers may utilize a wide variety of information associated with the firearm, scope, simulated ammunition, and/or simulated conditions (e.g., windage, elevation, etc.), as well as one or more ballistic models, to determine a ballistic path for the shot within the virtual environment. As a result of calculating a shot trajectory and/or a ballistic path, the accuracy of the simulated shot may be enhanced. Following the determination of information associated with a simulated shot, the image data output for presentation via the scope display may be updated.
  • System Overview
  • An example system 100 for facilitating firearms training will now be described illustratively with respect to FIG. 1. The system 100 may include, for example, a firearm 105, a scope 110 mounted onto the firearm 105, a display 115 associated with the scope 110, one or more motion and/or position tracking systems 120, and/or one or more controllers 125. In various embodiments, the motion tracking systems 120 may be configured to track or monitor motion, position, and/or orientation information associated with the firearm. Information collected by the motion tracking systems 120 may be utilized to determine a viewpoint associated with the firearm 105, and the determined viewpoint may be utilized to determine and/or generate image data (e.g., video data) to be rendered via the display 115.
  • With reference to FIG. 1, the firearm 105 may include any suitable firearm, modified firearm, or replica firearm suitable for use in a shooting or training simulation. For example, in certain embodiments, the firearm 105 may be a replica firearm (e.g., a replica or model rifle, pistol, shotgun, etc.) that has been modified to include one or more simulation components (e.g., a scope-mounted display, a scope-integrated display, one or more motion tracking systems, etc.). In other embodiments, the firearm 105 may be an actual firearm (e.g., rifle, pistol, shotgun, etc.) that has been modified to include or be otherwise associated with one or more simulation components. As a result of utilizing an actual firearm, a user may experience authentic trigger pressure, weapon weight, and/or weapon adjustments within a simulation environment. As desired, components may be incorporated into the actual firearm 105 and/or removably attached to the firearm 105. For example, simulation components and/or equipment may be integrated into a firearm to form a simulation firearm. As another example, a portion of the simulation components may be integrated into a firearm while other simulation components are attached to the firearm. As yet another example, simulation components and/or equipment may be attached to the outside of an actual weapon. For example, a user may attach simulation components to a personal firearm in order to utilize the personal firearm in a simulation environment. Indeed, a wide variety of firearm configurations may be utilized in accordance with various embodiments of the invention.
  • Additionally, in certain embodiments, a plurality of firearms may be utilized in association with a training system 100. Each of the firearms may include or be associated with respective simulation components (e.g., a display, one or more motion tracking systems, etc.). In the event that the system 100 includes one or more display devices situated external to the firearms, such as the one or more projection screens 130 illustrated in FIG. 1, the respective users associated with the firearms may be presented the same simulation or scenario by the external display devices. However, the individual displays associated with the respective firearm scopes may present the users with individual image data.
  • The scope 110 may be any suitable scope device, modified scope device, or replica scope device suitable for use in a shooting or training simulation. For example, in certain embodiments, the scope 110 may be a replica scope that has been modified to include or otherwise be associated with a display 115 and/or other simulation components. In other embodiments, the scope 11.0 may be an actual scope (i.e., a telescopic sight) that has been modified to include or otherwise be associated with a display and/or other simulation components. According to an aspect of the invention, a display 115 may be associated with the scope 110. In certain embodiments, the display 115 may be at least partially mounted within a housing of the scope 110. For example, a display 115 may be integrated into a scope 110 utilized in association with a modified simulation firearm. In other embodiments, such as embodiments in which a user utilizes a personal firearm, the display 115 may be removably attached to the scope. In yet other embodiments, a display 11.5 may be at least partially mounted within the housing of a scope 110 that is attached to an actual firearm when the firearm is utilized in conjunction with the simulation system 100.
  • The display 115 may be any suitable display device configured to present image data, such as video data, to a user of the firearm 105. Examples of suitable display devices include, but are not limited to, microdisplays, liquid crystal displays (“LCD's”), light emitting diode (“LED”) displays, organic light emitting diode (“OLED”) displays, active matrix OLED (“AMOLED”) displays, pico front or rear projection displays, stereoscopic displays, holographic displays, and/or cathode ray tube (“CRT”) displays. In certain embodiments, the display 115 may be configured to present high-definition imagery. Additionally, in certain embodiments, the display 115 may be at least partially mounted within the housing of the scope 110. In other embodiments, the display 115 may be removably attached to the scope 110. For example, the display may be attached to either end of the scope 110.
  • In certain embodiments, a video receiver 135 may be provided to route image and video signals to the display 115 for presentation to a user. For example, a video receiver 135 may receive image data from the one or more controllers 125, and the video receiver 135 may process and/or format the received image data. The processed and/or formatted image data may then be provided to the display 115. The video receiver 135 may be either a wired or wireless video receiver. For example, the video receiver 135 may be connected to the one or more controllers 125 via a video graphics array (“VGA”) cable, an Ethernet cable, a Universal Serial Bus (“USB”) cable, a mini-USB cable, a Firewire cable, a digital video interface (“DVI”) cable, a high definition multimedia interface (“HDMI”) cable, or other suitable cable or wired connection. As another example, the video receiver 135 may be connected to the one or more controllers 125 via a suitable Bluetooth connection, wireless Ethernet connection, Wi-Fi connection, or other wireless connection. In certain embodiments, one or more suitable communications interfaces (e.g., wired interfaces and/or wireless interfaces) 140 may facilitate communications between the video receiver 135 and the one or more controllers 125. Additionally, in certain embodiments, the video receiver 135 may be incorporated and/or integrated into the firearm 105 or the scope 110. In other embodiments, the video receiver 135 may be removably attached to the firearm 105.
  • As desired, an audio video (“AV”) receiver may be provided, and the AV receiver may process both audio and video data. The video data may be provided to the display 110, and the audio data may be provided to one or more suitable audio presentation devices (e.g., speakers, headphones, etc.). In certain embodiments, the AV receiver may receive separate audio and video signals. For example, separate audio and video signals may be received from the one or more controllers 125. As another example, a video signal may be received from the one or more controllers 125, and a separate audio signal may be received from one or more other controllers associated with the generation and presentation of a primary or main simulation scenario, such as a scenario presented via the one or more projection screens 130.
  • As desired, any number of suitable motion tracking systems 120 may be utilized in various embodiments of the invention. A motion tracking system 120 may be configured to monitor and/or track the motion, position, and/or orientation of the firearm 105 and/or the scope 110. In this regard, a viewpoint or perspective of the firearm 105 and/or the scope 110 may be determined, and image data may be determined and/or generated for presentation by the display 115. Examples of suitable motion tracking systems 120 that may be utilized include, but are not limited to, a laser tracking system, an electromagnetic tracking system, an optical tracking system, an infrared tracking system, an ultrasonic tracking system, an inertia tracking system, a mechanical tracking system, and/or a Global Positioning System (“GPS”) tracking system. In certain embodiments, a single motion tracking system may be utilized. In other embodiments, a plurality or combination of motion tracking systems may be utilized. In this regard, positional and/or orientation information for the firearm 105 and/or the scope 110 may be verified and/or accuracy may be improved.
  • A laser tracking system may include a laser emitter that is attached to or incorporated into the firearm 105 or the scope 110. For example, a laser emitter may be inserted into the barrel of the firearm 105. One or more laser detection devices, such as a laser camera or other laser detector, may also be attached to the firearm 105 and/or at one or more other positions suitable for viewing an area in which the laser emitter is pointed. For example, one or more laser detection devices may be utilized to identify an emitted laser that contacts and/or that is reflected off of the one or more projection screens 130. Based upon the capture of the emitted laser, positional and/or orientation data for the firearm 105 may be determined or calculated.
  • An electromagnetic tracking system may include one or more electromagnetic sources attached to the firearm 105 and one or more sensors utilized to monitor a calibrated area for position and motion of the electromagnetic sources. In this regard, positional and/or orientation data for the firearm 105 may be determined or calculated. In an optical tracking system, a video device may be calibrated to one or more unique optical markers attached to the firearm 105 and/or to one or more visual identifiers inherent to the firearm 105. Motion of the firearm 105 may be monitored by the video device, and information associated with the monitored motion may be translated into positional and/or orientation data for the firearm 105. With an infrared tracking system, a series of LEDs and/or reflective markers may be attached to the firearm 105. The LEDs may be utilized to emit infrared signals that are reflected by the markers. One or more infrared cameras may be utilized to monitor a field of view encompassing the firearm, such as a simulation area. As the firearm 105 is moved, the infrared camera may identify movement of the LEDs and/or reflective markers, and information associated with the identified movement may be translated into positional and/or orientation data for the firearm 105. In certain embodiments, movement of the LEDs and/or reflective markers may be identified relative to one another in order to provide a perspective for the firearm 105.
  • An ultrasonic tracking system may include one or more ultrasonic transmitters that are attached to the firearm 105. One or more audio sensors situated in predetermined positions may be utilized to measure the time-of-flight and/or the phase of the ultrasonic sound emitted by the ultrasonic transmitters. The measurements data may then be processed in order to determine distances between the sensors and the firearm, and the distances may be utilized to determine positional and/or orientation data for the firearm 105. With an inertia tracking system, one or more inertia sensors including gyro meters and/or accelerometers may be attached to the firearm 105. Motion or movement of the firearm 105 may be measured as inertia and translated into positional and/or orientation data for the firearm 105. A mechanical tracking system may include one or more mechanical devices and/or mechanical sensors that are attached to the firearm 105 such that movement of the firearm 105 directly affects the mechanical devices. Movement of the gun may be translated into movement of portions of the mechanical devices and may be translated into positional and/or orientation data. With a GPS tracking system, one or more GPS receivers may be attached to the firearm 105. A GPS receiver may precisely measure the timing of signals output by orbital GPS satellites. As the firearm 105 moves, the GAS receiver continually calculates positional information in real-time and tracks movement of the firearm 105. The movement data may then be translated into positional and/or orientation information for the firearm 105.
  • According to an aspect of the invention, movement and/or motion of the firearm 105 may be tracked along any number of axes utilizing two to six degrees of freedom. For example, motion may be tracked along an x, y, and/or z axis. Additionally, two, three, four, five, or six degrees of freedom may be utilized to track or monitor motion of the firearm 105. In this regard, motion may be selectively tracked as the firearm is moved forward and backward (e.g., closer to or further away from a projection screen), left or right, and/or up or down. Additionally, yaw, pitch, and/or roll of the firearm 105 may be tracked. The tracking of the firearm may be utilized to determine a viewpoint and/or perspective of the firearm in order to determine image data for presentation via the display 115.
  • As desired in certain embodiments, a wide variety of other components may be incorporated into, attached to, and/or utilized in conjunction with the firearm 105. For example, one or more communications interfaces 140 may be incorporated into and/or attached to the firearm 105 or the scope 110. The communications interfaces may include wired and/or wireless communications devices (e.g., transceivers, communications cards, etc.) that facilitate communication with the controllers 125. In this regard, data and/or information collected by simulation components attached to the firearm 105 (e.g., the motion tracking systems 120, a camera, a trigger recognition device, a keypad, etc.) may be communicated to the controllers 125. Additionally, data output by the controllers 125 (e.g., image data, calibration data etc.) may be received and distributed to the various simulation components.
  • In certain embodiments, a camera 145 may be integrated into or incorporated into the firearm 105 or the scope 110. The camera 145 may be utilized to detect and/or monitor grid patterns projected onto or output by an external display surface, such as the one or more projection screens 130, The data captured by the camera 145 may be processed in order to determine an orientation and/or perspective of the firearm 105. In this regard, the motion tracking of the firearm 105 may be verified. Additionally, a determination may be made as to whether the firearm 105 is positioned correctly.
  • Additionally, one or more trigger mechanisms 150 and/or trigger detection devices may be incorporated into, inserted into, and/or attached to the firearm 105. These mechanisms and/or devices may facilitate the detection of a trigger pull or a trigger actuation. Once a trigger actuation has been detected, an indication of the detected actuation may be communicated to the controllers 125, and a shot event may be registered. A wide variety of trigger mechanisms 150 and/or trigger detection devices may be utilized as desired in various embodiments of the invention. For example, a firearm 105 may be modified to include a trigger mechanism that includes one or more actuation sensors. As another example, an inert or dummy round may be inserted into a firearm 105, and the dummy round may include one or more suitable electrical components that detect actuation of the firing pin. An inert round is explained in greater detail below with reference to FIGS. 2A and 2B. The trigger mechanisms and/or trigger detection devices described above are provided by way of example only, and other suitable devices may be utilized as desired in various embodiments of the invention.
  • A wide variety of other sensors may be incorporated into and/or attached to the firearm 105 as desired in various embodiments. For example, one or more sensors may be attached to the scope 110 in order to detect user adjustment made to the scope, such as windage adjustments, elevation adjustments, focus adjustments, zoom adjustments, and/or mode adjustments (e.g., a standard mode, a thermal imaging mode, a night vision mode, etc.). As another example, one or more sensors may be utilized to monitor the removal and/or insertion of magazines. As desired, information collected by the sensors may be communicated to the controllers 125 and utilized during the presentation of audio and/or image data during a simulation.
  • With continued reference to FIG. 1, any number of controllers 125 may be provided. In certain embodiments, a controller 125 may be situated or positioned external to the firearm 105 and/or the scope 110. In other embodiments, a controller 125 may be incorporated into and/or attached to the firearm 105 and/or the scope 110. Each controller 125 may be a suitable processor-driven device that facilitates the provision of a firearms simulation. For example, a controller 125 may be a suitable processor-driven device that receives motion information associated with the firearm 105, determines a viewpoint of the firearm 105, generates or determines image data for presentation via the display 115, and/or outputs the image data for presentation via the display 115. As such, the controller 125 may include any number of computing devices, such as a personal computer, a digital assistant, a personal digital assistant, a digital tablet, an Internet appliance, an application-specific circuit, a microcontroller, a minicomputer, or any other processor-based device. The execution of suitable computer-implemented instructions or computer-executable instructions by the controller 125 may form a special purpose computer or other particular machine that is operable to facilitate the provision of a firearms simulation.
  • With reference to FIG. 1, a controller 125 may include one or more processors 152, one or more memory devices 154, one or more network or communications interfaces 156, and/or one or more input/output (“I/O”) interfaces 158. The processors 152 may be configured to execute any number of software applications and/or computer-readable or computer-executable instructions. The memory devices 154 may include any number of suitable memory devices, such as caches, read-only memory devices, random access memory devices, flash memory devices, magnetic storage devices, removable storage devices (e.g., memory cards, memory sticks, etc.), etc. The memory devices 154 may include internal memory devices and/or external memory devices in communication with the controller 1.25. The memory devices 154 may store data, executable instructions, and/or various program modules utilized by the processors 152, such as data files 160, an operating system (“OS”) 162, a tracking module 164, a video module 166, and/or a ballistics module 168.
  • The data files 160 may include any suitable data that facilitates the operation of the controller 125 and/or the provision of a firearms simulation. For example, the data files 160 may include, but are not limited to, received motion and/or movement data for one or more firearms 105, calibration information for a simulation environment, information associated with types of rounds or ammunition to be simulated, information associated with desired scope settings, information associated with desired simulation conditions (e.g., wind, etc.), one or more ballistic models, stored image and/or audio data, and/or information utilized to generate and/or format image and/or audio data for presentation during a simulation.
  • The OS 162 may be a suitable software module that controls the general operation of the controller 125 and/or the execution of other software modules, for example, the tracking module 164, the video module 166, and/or the ballistics module 168. The tracking module 164 may be a suitable software module or application that facilitates the determination of a viewpoint of the firearm 105. The viewpoint of the firearm 105 may then be utilized to determine video to be presented via the display 115. As desired, the viewpoint of the firearm 105 may take a wide variety of position, orientation, and/or perspective information into account, as well as various settings associated with the scope 110 (e.g., a magnification setting, an operational mode, etc.).
  • In operation, the tracking module 164 may receive measurements and/or tracking data from the one or more motion tracking systems 120, and the tracking module 164 may utilize at least a portion of the received data to determine a viewpoint of the firearm 105. For example, received measurements data may be translated into positional and/or orientation information for the firearm 105. The positional and/or orientation information may then be utilized to determine a viewpoint for the firearm 105. For example, an area of the one or more projection screens 130 at which the firearm 105 is aimed or pointed may be determined. In certain embodiments, various tilts and/or angles of the firearm (e.g., pitch, yaw, roll) may be taken into consideration when determining the viewpoint. Additionally, in certain embodiments, a distance between the firearm 105 and a relevant projection screen 130 may be taken into consideration when determining the viewpoint.
  • In certain embodiments, the tracking module 164 may receive tracking data from a plurality of motion tracking systems 120, and the tracking module 164 may process the respective data to verify the positional and/or orientation information for the firearm 105. As desired, positional and/or orientation information may be averaged or otherwise processed in the event that tracking data is received from a plurality of sources. Additionally, in certain embodiments, grid pattern information may be received from the camera 145 and utilized to verify a viewpoint of the firearm 105.
  • In certain embodiments, such as embodiments in which a plurality of firearms are utilized in conjunction with a simulation, the tracking module 164 may correct or adjust respective positions of the one or more firearms. In this regard, image data that is determined or generated for display via the firearm scopes may correspond to the perspective illustrated via the external projection screens 130, which is the same for all users within the simulation environment. In other words, prior to determining a viewpoint of a firearm 105, the position of the firearm 105 may be corrected to a calibration point for the simulation environment. In this regard, the image data presented via the display 115 may correspond to that presented to the user via the projection screens 130, regardless of the user's movement within a simulation environment. In other embodiments, such as embodiments with a single user or embodiments in which each user has a personal external display (e.g., a personal monitor, a headset, etc.), a simulation scenario illustrated on the one or more external displays may be modified as the user moves within a simulation environment.
  • The video module 166 may be a suitable software module or application configured to determine and/or generate image data for presentation via the display 115. For example, the video module 166 may receive viewpoint information for the firearm 105, and the video module 166 may utilize the viewpoint information to generate image data for communication to the video receiver 135 and/or the display 115. As desired, the video module 166 may additionally utilize one or more settings associated with the scope (e.g., a magnification, a display mode, etc.) and/or an indication of a shot event to determine or generate image data for presentation. A wide variety of image data may be generated by the video module 166, including simulation scenario imagery, enhanced or enlarged imagery corresponding to a portion or section of the imagery displayed on the projection screens 130, imagery associated with a simulated view through the scope 110 (e.g., a reticle, etc.), and/or imagery associated with a simulated shot. In certain embodiments, the video module 166 may also include an audio component configured to determine or generate audio for presentation to a user of the firearm 105.
  • The ballistics module 168 may be a suitable software module or application configured to calculate or determine a simulated trajectory of a shot within a virtual environment. Once a shot event has been identified, information associated with the viewpoint and/or orientation of the firearm 105 may be provided to the ballistics module 168, and the ballistics module 168 may calculate a trajectory for the shot. As desired, the ballistics module 168 may utilize a wide variety of other information to calculate a trajectory for the shot. For example, information associated with the firearm 105 (e.g., caliber information, etc.), information associated with a simulated ammunition, elevation information and/or wind information may be provided to any number of suitable ballistic models to calculate or determine a trajectory for the simulated shot. Results of the simulated shot may then be provided to the video module 166 to facilitate the generation of image data associated with the simulated shot. As a result of calculating shots by the ballistics module 168 within a virtual environment, shot accuracy may be improved or enhanced. In conventional training systems, such as systems that detect a shot by detecting a laser projected onto a two-dimensional surface, the size of the laser point or other shot may be bigger than the target aimed at, thereby reducing accuracy and ballistic fidelity. By simulating shots directly within a virtual environment, external shot detection mechanisms may be bypassed, and overall accuracy may be improved.
  • A wide variety of suitable operations may be performed by the simulation components associated with a firearm 105 and/or by the one or more controllers 125 to facilitate provision of a training simulation. The various operations described above for a controller 125 are provided by way of example only. Another example of the operations that may be performed by the controllers 125 is described in greater detail below with reference to FIG. 8. Additionally, an example of the operations that may be performed by the simulation components associated with the firearm 105 is described in greater detail below with reference to FIG. 7.
  • With continued reference to the controller 125 of FIG. 1, the one or more I/O interfaces 158 may facilitate communication between the controller 125 and one or more input/output devices, for example, one or more user interface devices, such as a display, keypad, mouse, pointing device, control panel, touch screen display, remote control, microphone, speaker, etc., that facilitate user interaction with the controller 125. In this regard, user commands may be received by the controller 125, and a wide variety of data may be output for presentation to a user. The one or more network interfaces 156 may facilitate connection of the controller 125 to one or more suitable wired and/or wireless networks 145, for example, a local area network, a Wi-Fi enabled network, a Bluetooth-enabled network, a radio frequency network, the Internet, etc. Communications between various components of the system 100 may be facilitated via the networks 145.
  • With continued reference to FIG. 1, one or more projection screens 130 and/or other external display devices may be utilized in conjunction with the training system 100. For example, one or more projection screens 130 may be utilized to present a simulation scenario (e.g., a military training exercise, a law enforcement training exercise, a hunting scenario, etc.) to any number of users associated with respective firearms. A wide variety of projection screen arrangements may be utilized in various embodiments of the invention. For example, a plurality of screens may be placed next to each other to enlarge a simulation viewing area. As another example, the projections screens may positioned in a curve to create up to a 360 degree simulation viewing area. Indeed, projection screens may even be positioned in a half dome or full dome configuration. As desired, one or more projectors 170 or other image generation devices may be configured to output images and/or video presented via the projection screens 130.
  • Additionally, in certain embodiments, one or more simulation controllers 175 may be provided. A simulation controller 175 may be a suitable processor-driven device configured to control the presentation of a simulation via the one or more projection screens 130. As desired, a simulation controller 175 may include components similar to those described above for the controller 125. For example, a simulation controller 175 may include one or more processors and/or memory devices, and simulation software may be executed by the processors to facilitate the generation of image data to be. displayed via the projection screens 130. In certain embodiments, a simulation controller 175 may be separate from one or more controllers 125 associated with individual firearms. In other embodiments, the simulation controller 175 may be combined with one or more other controllers 125.
  • Those of ordinary skill in the art will appreciate that the system 100 shown in and described with respect to FIG. 1 is provided by way of example only. Numerous other operating environments, system architectures, and device configurations are possible. Other system embodiments can include fewer or greater numbers of components and may incorporate some or all of the functionality described with respect to the system components shown in FIG. 1.
  • As set forth above, a wide variety of suitable techniques may be utilized as desired to modify firearms for use in a simulation environment. FIGS. 2-5 illustrate a few example modified firearms that may be utilized. FIG. 2A is an illustration of an actual firearm 200 that has been modified to include attached simulation components. The firearm 200 may be any suitable firearm onto which a scope 205 has been mounted. With reference to FIG. 2A, the scope 205 may be an actual scope mounted to the firearm 200. A scope attachment 210 may be connected (e.g., removably connected) to one end of the scope 205 to facilitate the presentation of image data to a user. The scope attachment 210 may include a suitable visual display unit configured to present image data. A visual display unit may include, for example, a lens, a digital display micro-screen or other suitable display, a control board, and/or a communications interface.
  • With continued reference to FIG. 2A, one or more tracking sensors 215 may be attached to the firearm 200 to facilitate motion tracking of the firearm 200. Additionally, an inert round 220 may be inserted into the firearm 200 to facilitate the detection of a trigger pull event. FIG. 2B is an exploded illustration of the firearm 200 shown in FIG. 2A. Also shown in FIG. 2B is the interior 225 of the inert round 220. The interior 225 may be a spring-mounted metal bolt. A circuit 230, such as an approximately five (5) volt circuit, may be connected to the inert round 220 and grounded to the firearm 200. A trigger pull even may lead to the inert round being contacted by the firing pin of the firearm 200, thereby causing the circuit 230 to detect a shot being fired.
  • Although FIGS. 2A-2B illustrate an actual firearm, replica or model firearms may be modified in a similar manner for use in a simulation system.
  • FIG. 3 illustrates a block diagram of an example scope 300 that may be mounted on a firearm in accordance with various embodiments of the invention. The scope 300 may include any number of incorporated, integrated, and/or attached simulation components. For example, the scope 300 may include a suitable display device 305, a video receiver 310, and/or one or more communications interfaces 315. Each of these components may be similar to the components described in greater detail above with reference to FIG. 1. These components may facilitate the receipt of image data output by a controller, and the presentation of the image data to a user.
  • As desired, the communications interfaces 315 associated with the scope 300 may also be utilized to facilitate communication of measurements data to one or more controllers. For example, the communications interfaces 315 may be in communication with a trigger mechanism 320, a motion tracking system 325, and/or a camera 330. In this regard, shot indication data, motion data, and/or grid pattern data may be communicated to one or more controllers for processing. Although FIG. 3 illustrates a motion tracking system 325 and camera 330 that have been integrated into the scope 300, one or more of these components may be external to the scope 300.
  • FIG. 4 is an illustration of an actual firearm 400 onto which a replica scope 405 has been mounted to facilitate the use of the firearm 400 within a simulation environment. The replica scope 405 may include components similar to those described above with reference to FIG. 3. With additional reference to FIG. 4, in certain embodiments, one or more tracking sensors 410 may be attached to the firearm 400 to facilitate motion tracking. Additionally, an inert round 415 may be inserted into the firearm to facilitate detection of shot events.
  • FIG. 5 is an illustration of a firearm 500 into which simulation components have been integrated. The firearm 500 illustrated in FIG. 5 may be a replica firearm or an actual firearm that has been modified to include integrated simulation components. For example, the firearm 500 may be modified to include a trigger mechanism, a tracking system, and/or a communications interface. Similarly, the scope may be modified to include a display device and a video receiver.
  • Indeed, a wide variety of different firearm configurations may be utilized in accordance with various embodiments of the invention. These configurations should not be limited to those described in FIGS. 2-5. Additionally, a wide variety of different types of firearms and/or other weapons may be utilized in conjunction with the embodiments of the invention. Embodiments of the invention are applicable to a myriad of weapon types that use an optical aiming system, such as rifles, handguns, missile launchers, etc.
  • FIG. 6 is a diagram of a simulation environment 600 that may be utilized in accordance with certain embodiments of the invention. The simulation environment 600 may include any number of displays 605, such as projection screens, that facilitate presentation of simulation scenarios. Although three projection screens are illustrated in FIG. 6, any number of suitable displays may be utilized. For example, a relatively circular simulation environment may be provided in which displays surround users in three hundred and sixty degrees. Additionally, as desired, any number of firearms 610 a-n may be utilized in conjunction with the simulation environment 600. Each of the firearms 610 a-n may be associated with a respective scope-integrated or scope-mounted display. Thus, users may view a simulation scenario within the simulation environment 600 via the external displays 605. Each user may then use a scope-mounted display associated with his/her firearm to view image data determined based upon the viewpoint of the firearm.
  • As shown in FIG. 6, a perspective may be predetermined or assumed for the external displays 605. For example, a calibration point 615 may be situated at a relatively central point for viewing the external displays 605. As desired, the firearms may be calibrated for use with the simulation system 600 at the calibration point 615.
  • During a simulation, a user may move a firearm, such as firearm 610 a, within the simulation environment 600. As the firearm is moved, the firearm 610 a may be pointed at different portions of the external displays 605. Accordingly, the viewpoint of the firearm 610 a may change as the firearm 610 a is moved. During the generation or determination of image data to be presented via a scope-mounted display associated with the firearm 610 a, the position of the firearm 610 a within the simulation environment 600 may be autocorrected to the calibration point 615. For example, an area or portion of the external displays 605 at which the firearm 610 a is aimed may be determined. A viewpoint of that area or portion from the perspective of the calibration point may then be determined as the viewpoint of the firearm 610 a. In other words, regardless of the user's position within the simulation environment, it may be assumed that the user is positioned at the calibration point 615. Accordingly, the imagery illustrated via the scope-mounted display may correlate or correspond to the imagery presented to the user via the external displays 605.
  • Operational Overview
  • FIG. 7 is a flow diagram of an example method 700 for presenting image data via a display associated with a scope mounted onto a firearm, according to an example embodiment of the invention. In certain embodiments, the method 700 may be performed by a suitable firearm, such as the firearm 105 illustrated in FIG. 1. The method 700 may begin at block 705.
  • At block 705, communication may be initiated with a host controller, such as a host computer. In this regard, motion tracking information may be communicated to the host controller, and the image data may be received from the host controller for presentation via the display associated with the scope.
  • At block 710, interaction with one or more tracking systems (e.g., a laser tracking system, an electromagnetic tracking system, an optical tracking system, etc.) may be initiated. Each tracking system may be configured to track motion and/or movement of the firearm 105 within a simulation environment. In certain embodiments, a tracking system may be configured to determine position and/or orientation information for the firearm 105. For example, as illustrated in optional block 715, position and/or orientation information for the firearm 105 may be determined, and the position and/or orientation information may be communicated to the host controller at block 720.
  • In other embodiments; tracking information collected by a tracking system may be provided or communicated to the host controller at block 720, and the host controller may determine position and/or orientation information for the firearm 105. Regardless of the component or system that determines the position and/or orientation information for the firearm 105, a viewpoint of the firearm 105 or the firearm scope may be determined by the host controller, and the host controller may generate image data in accordance with the determined viewpoint. The image data may be output by the host controller for receipt by the firearm 105, and the firearm 105 may receive the image data at block 725. The received image data, such as a received image feed, may then be output by the display for presentation to the user at block 730. Accordingly, when a user views the scope associated with the firearm 105, the user may view the image data output by the host controller.
  • At block 735, a determination may be made as to whether the image is calibrated. In other words, a determination may be made as to whether the firearm 105 has been properly calibrated within the simulation system and/or the virtual environment. If it is determined at block 735 that the image is not calibrated, then operations may continue at block 740. At block 740, the firearm 105 and/or the position of the firearm 105 may be calibrated within the virtual environment. For example, a user may be prompted to utilize a suitable calibration routine (e.g., aiming the firearm 105 at the four corners of an external display, etc.) to calibrate the firearm 105. Operations may then continue at block 725 described above.
  • If, however, it is determined at block 735 that the image is calibrated, then operations may continue at block 745. At block 745, a determination may be made as to whether a trigger associated with the firearm 105 has been pulled or actuated by a user. In other words, a determination may be made as to whether a shot has occurred. As described in greater detail above with reference to FIG. 1, a wide variety of suitable trigger mechanisms and/or trigger detection devices (e.g., an inert round, etc.) may be utilized to detect a trigger actuation. If it is determined at block 745 that the trigger has not been actuated or pulled, then operations may continue at block 765 described in greater detail below. If, however, it is determined at block 745 that the trigger has been actuated or pulled, then operations may continue at block 750.
  • At block 750, a shot indication may be communicated from the firearm 105 to the host controller. The shot indication may be processed by the host controller to simulate a shot within a virtual environment. For example, one or more ballistic models may be utilized to determine a trajectory of a shot within the virtual environment. Image data (e.g., an image feed) may then be updated by the host controller to reflect the shot, and the updated image data may be output by the host controller for receipt by the firearm 105. The firearm 105 may receive the updated image data at block 755 and present information associated with a virtual shot via the display at block 760. Operations may then continue at block 765.
  • At block 765, a determination may be made as to whether a position and/or orientation of the firearm 105 has changed. For example, a determination may be made as to whether the firearm 105 has been moved or whether a tracking system has detected movement of the firearm 105. If it is determined at block 765 that the position and/or orientation of the firearm 105 has not changed, then operations may continue at block 730, and image data may continue to be output. If, however, it is determined at block 765 that the position and/or orientation of the firearm 105 has changed, then operations may continue at block 715 (or block 720). In this regard, a new viewpoint of the firearm 105 may be determined, and the image data output by the display may be updated to reflect the new viewpoint.
  • The method 700 may end following the completion of a simulation scenario in which the firearm 105 is utilized.
  • FIG. 8 is a flow diagram of an example method 800 for determining image data to be presented via a display associated with a scope mounted onto a firearm, according to an example embodiment of the invention. In certain embodiments, the method 800 may be performed by one or more suitable controllers associated with a simulation system, such as the controllers 125 illustrated in FIG. 1. The method 800 may begin at block 805.
  • At block 805, the controller 125 may initiate communication with a firearm and/or a scope, such as the firearm 105 and/or the scope 110 illustrated in FIG. 1. In this regard, the controller 125 may receive a wide variety of data and/or indications from the firearm 105, such as indications associated with shot events or trigger actuations and/or information associated with scope settings. A communications channel may also be established for communicating image data to the firearm 105 and/or the scope 110.
  • At block 810, one or more firearm and/or scope parameters may be determined. In certain embodiments, the one or more parameters may be determined or identified within suitable ballistics simulation software, such as the ballistics module 168 illustrated in FIG. 1. A wide variety of different firearm and/or scope parameters may be determined as desired in various embodiments of the invention, such as a type associated with the firearm 105 (e.g., a caliber, etc.), a ballistic profile associated with the firearm 105, a type of ammunition to be simulated, a ballistic profile associated with the ammunition, a zoom or magnification for the scope 110, a mode of operation for the scope 110 (e.g., normal operation, an infrared mode, a night vision mode, etc.), and/or windage and/or elevation adjustments for the scope 110, etc. In this regard, the trajectory of simulated shots may be determined in a relatively accurate manner.
  • At block 815, the controller 125 may initiate communication with one or more suitable tracking systems associated with the firearm 105 and/or the scope 110, such as the motion tracking system 120 illustrated in FIG. 1. In this regard, the controller 125 may receive motion, position, and/or orientation data associated with the firearm 105. At block 820, position, orientation, and/or viewpoint information for the firearm 105 may be determined. For example, tracking data received from the motion tracking systems 120 may be translated into position and/or orientation data for the firearm 105. A viewpoint of the firearm 105 within the virtual environment may then be determined.
  • At block 825, which may be optional in certain embodiments of the invention, a position of the firearm 105 within the virtual environment may be corrected or adjusted. For example, the position of the firearm 105 may be set to a predetermined position, such as a predetermined calibration point. In this regard, a determined viewpoint for the firearm 105 may correspond with the viewpoint presented via one or more external (i.e., non-firearm-mounted) displays.
  • At block 830, an image feed for display via a scope-mounted display may be determined and/or generated based at least in part upon the determined viewpoint for the firearm 105. The image feed may then be output at block 835 for receipt by the firearm 105 and presentation to the user via the scope-mounted display. At block 840, a determination may be made as to whether the image communicated to the firearm 105 is calibrated. In other words, a determination may be made as to whether the firearm 105 has been properly calibrated within the simulation system and/or the virtual environment. If it is determined at block 840 that the image is not calibrated, then operations may continue at block 845. At block 845, the firearm 105 and/or the position of the firearm 105 may be calibrated within the virtual environment. For example, a user may be prompted to utilize a suitable calibration routine (e.g., aiming the firearm 105 at the four corners of an external display, etc.) to calibrate the firearm 105. Operations may then continue at block 830 described above.
  • If, however, it is determined at block 840 that the image is calibrated, then operations may continue at block 850. At block 850, a determination may be made as to whether a shot event has occurred. For example, a determination may be made as to whether a shot indication has been received from the firearm 105. If it is determined at block 850 that a shot event has not occurred, then operations may continue at block 865 described in greater detail below. lf, however, it is determined at block 850 that a shot event has occurred, then operations may continue at block 855.
  • At block 855, the shot event may be registered within the ballistic simulation software, and a trajectory of the shot may be calculated or determined. As desired, a wide variety of ballistic factors, such as weather conditions, elevation, ammunition type, firearm type, etc., may be taken into consideration during the calculation or determination of a shot trajectory. Once a shot trajectory has been calculated, operations may continue at block 860, and the image feed may be adjusted or modified to correspond with the registered shot. Image information associated with the simulated shot may then be output for receipt via the firearm 105 and presentation via the scope-mounted display. Operations may then continue at block 865.
  • At block 865, a determination may be made as to whether a position and/or orientation of the firearm 105 has changed. For example, a determination may be made as to whether motion or movement information has been received from one or more tracking systems. As another example, a determination may be made as to whether updated position and/or orientation information has been received. If it is determined at block 865 that the position and/or orientation of the firearm 105 has not changed, then operations may continue at block 835, and image data may continue to be output for receipt by the firearm 105. If, however, it is determined at block 865 that the position and/or orientation of the firearm 105 has changed, then operations may continue at block 820, and a new viewpoint for the firearm 105 may be determined. In this regard, the image data output for display by the scope may be updated to reflect the new viewpoint.
  • The method 800 may end following the completion of a simulation scenario associated with the controller 125.
  • The operations described and shown in the methods 700 and 800 of FIGS. 7-8 may be carried out or performed in any suitable order as desired in various embodiments of the invention. Additionally, in certain embodiments, at least a portion of the operations may be carried out in parallel. Furthermore, in certain embodiments, less than or more than the operations described in FIGS. 7-8 may be performed.
  • Various block and/or flow diagrams of systems, methods, apparatus, and/or computer program products according to example embodiments of the invention are described above. It will be understood that one or more blocks of the block diagrams and flow diagrams, and combinations of blocks in the block diagrams and flow diagrams, respectively, can be implemented by computer-executable program instructions. Likewise, some blocks of the block diagrams and flow diagrams may not necessarily need to be performed in the order presented, or may not necessarily need to be performed at all, according to some embodiments of the invention.
  • These computer-executable program instructions may be loaded onto a special purpose computer or other particular machine, a processor, or other programmable data processing apparatus to produce a particular machine, such that the instructions that execute on the computer, processor, or other programmable data processing apparatus create means for implementing one or more functions specified in the flow diagram block or blocks. These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means that implement one or more functions specified in the flow diagram block or blocks. As an example, embodiments of the invention may provide for a computer program product, comprising a computer-usable medium having a computer-readable program code or program instructions embodied therein, said computer-readable program code adapted to be executed to implement one or more functions specified in the flow diagram block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational elements or steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process such that the instructions that execute on the computer or other programmable apparatus provide elements or steps for implementing the functions specified in the flow diagram block or blocks.
  • Accordingly, blocks of the block diagrams and flow diagrams support combinations of means for performing the specified functions, combinations of elements or steps for performing the specified functions and program instruction means for performing the specified functions. It will also be understood that each block of the block diagrams and flow diagrams, and combinations of blocks in the block diagrams and flow diagrams, can be implemented by special purpose, hardware-based computer systems that perform the specified functions, elements or steps, or combinations of special purpose hardware and computer instructions.
  • Many modifications and other embodiments of the invention set forth herein will be apparent having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. Therefore, it is to be understood that the invention is not to be limited to the specific embodiments disclosed and that modifications and other embodiments are intended to be included within the scope of the appended claims. Although specific terms are employed herein, they are used in a generic and descriptive sense only and not for purposes of limitation.

Claims (20)

1. A firearm training system, comprising:
a firearm;
a scope mounted to the firearm, the scope comprising a display;
a tracking system configured to collect information associated with a position and an orientation of the firearm; and
at least one controller configured to (i) determine, based at least in part upon the information collected by the tracking system, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.
2. The firearm training system of claim 1, wherein the at least one controller is further configured to adjust, prior to determining the image data to be rendered, the position of the firearm within a virtual environment.
3. The firearm training system of claim 2, wherein the display comprises a first display, and further comprising:
a second display external to the firearm, the second display configured to render a representation of the virtual environment;
wherein the at least one controller is configured to adjust the position of the firearm to a calibration point associated with the second display.
4. The firearm training system of claim 1, wherein the tracking system comprises a laser tracking system comprising:
a laser emitter attached to the firearm; and
at least one laser detector configured to detect a laser emitted by the laser emitter.
5. The firearm training system of claim 1, wherein the tracking system is configured to track motion along one, two, or three axes with two to six degrees of freedom.
6. The firearm training system of claim 1, wherein the tracking system comprises one of (i) an electromagnetic tracking system, (ii) an optical tracking system, (iii) an infrared tracking system, (iv) an ultrasonic tracking system, (v) an inertia tracking system, (vi) a mechanical tracking system, or (vii) a global positioning system tracking system.
7. The firearm training system of claim 1, wherein the display is mounted at least partially within a housing of the scope.
8. The firearm training system of claim 1, wherein the display comprises one of (i) a microdisplay, (ii) a liquid crystal display, (iii) a light emitting diode display, or (iv) an organic light emitting diode display.
9. The firearm training system of claim 1, further comprising:
a trigger assembly associated with the firearm, the trigger assembly configured to detect a user actuation of a trigger,
wherein the at least one controller is further configured to (i) receive an indication of the trigger actuation, (ii) calculate a trajectory of a shot within a virtual environment, and (iii) direct output of visual information associated with the shot for receipt by the display.
10. The firearm training system of claim 1, further comprising:
a camera attached to the firearm and configured to read one or more grid patterns associated with a surface external to the firearm,
wherein the at least one controller is further configured to (i) receive grid pattern information from the camera and (ii) verify the determined viewpoint of the firearm utilizing at least a portion of the received grid pattern information.
11. A system for modifying a firearm for use in a simulation environment, the system comprising:
a display configured to be attached to a scope associated with the firearm;
a tracking component configured to be attached to the firearm, the tracking component further configured to collect information associated with a position and an orientation of the firearm; and
at least one controller configured to (i) determine, based at least in part upon the information collected by the tracking component, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.
12. The system of claim 11, wherein the at least one controller is further configured to adjust, prior to determining the image data to be rendered, the position of the firearm within a virtual environment.
13. The system of claim 1 I, wherein the tracking system comprises a laser tracking system comprising:
a laser emitter attached to the firearm; and
at least one laser detector configured to detect a laser emitted by the laser emitter.
14. The system of claim 11, wherein the display comprises one of (i) a microdisplay, (ii) a liquid crystal display, (iii) a light emitting diode display, or (iv) an organic light emitting diode display.
15. The system of claim 11, further comprising:
a trigger detection mechanism configured to be attached to the firearm, wherein the trigger detection mechanism detects a user actuation of a trigger associated with the firearm, and
wherein the at least one controller is further configured to (i) receive an indication of the trigger actuation, (ii) calculate a trajectory of a shot within a virtual environment, and (iii) direct output of visual information associated with the shot for receipt by the display.
16. A firearm scope, comprising:
a housing configured to be mounted onto a firearm;
a display mounted at least partially within the housing; and
a video receiver in communication with the display and configured to output video to be rendered by the display,
wherein the video receiver receives the video from at least one controller configured to (i) determine a viewpoint of the firearm, (ii) adjust a position of the firearm within a virtual environment, (iii) determine the video to be rendered, and (iv) direct communication of the video to the video receiver.
17. The firearm scope of claim 16, further comprising:
a laser tracking system comprising a laser emitter and a laser detector configured to detect a laser emitted by the laser emitter,
wherein the at least one controller determines the viewpoint of the firearm based at least in part upon tracking information received from the laser tracking system.
18. A method for providing a firearm simulation, the method comprising:
tracking, via a tracking component attached to a firearm, motion of the firearm;
determining, by a controller comprising one or more computers and based upon information collected by the tracking component, a viewpoint of the firearm;
determining, by the controller based at least in part upon the determined viewpoint, image data to be rendered by a display associated with a scope mounted onto the firearm;
outputting, by the controller for receipt by the display, the determined image data; and
presenting the image data by the display.
19. The method of claim 18, further comprising:
adjusting, by the controller and prior to determining the image data to be rendered, a position of the firearm within a virtual environment.
20. The method of claim 18, wherein tracking motion of the firearm comprises tracking motion utilizing a laser tracking component.
US13/035,321 2010-02-25 2011-02-25 Firearm training systems and methods of using the same Abandoned US20110207089A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/035,321 US20110207089A1 (en) 2010-02-25 2011-02-25 Firearm training systems and methods of using the same

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US30816910P 2010-02-25 2010-02-25
US13/035,321 US20110207089A1 (en) 2010-02-25 2011-02-25 Firearm training systems and methods of using the same

Publications (1)

Publication Number Publication Date
US20110207089A1 true US20110207089A1 (en) 2011-08-25

Family

ID=44476805

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/035,321 Abandoned US20110207089A1 (en) 2010-02-25 2011-02-25 Firearm training systems and methods of using the same

Country Status (1)

Country Link
US (1) US20110207089A1 (en)

Cited By (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110315767A1 (en) * 2010-06-28 2011-12-29 Lowrance John L Automatically adjustable gun sight
US20120115110A1 (en) * 2010-11-10 2012-05-10 Christian Patrick Wade Method and tool for firearm sight training
US20130040268A1 (en) * 2010-04-23 2013-02-14 Tacktech Pty (Ltd) Simulated shooting device and system
US20130228618A1 (en) * 2012-03-05 2013-09-05 James Allen Millett Dscope aiming device
US20130288205A1 (en) * 2012-04-30 2013-10-31 Trackingpoint, Inc. Rifle Scope and Method of Providing Embedded Training
US20130344461A1 (en) * 2012-06-26 2013-12-26 Selso Tello Universal Firearm Marksmanship System
USRE44786E1 (en) 2005-12-22 2014-03-04 Force Science Institute Ltd. System and method for monitoring handling of a firearm or other trigger-based device
US20140182187A1 (en) * 2012-12-31 2014-07-03 Trackingpoint, Inc. Software-Extensible Gun Scope and Method
US20140272807A1 (en) * 2013-03-15 2014-09-18 Kenneth W. Guenther Interactive system and method for shooting and target tracking for self-improvement and training
WO2014145079A1 (en) * 2013-03-15 2014-09-18 Jacob Stewart Schmehl Firearm accessory
US8857714B2 (en) * 2012-03-15 2014-10-14 Flir Systems, Inc. Ballistic sight system
US8881981B2 (en) 2012-03-05 2014-11-11 James A. Millett Digital targeting scope apparatus
US8950102B1 (en) 2012-04-13 2015-02-10 The Board Of Trustees Of The University Of Alabama Scope correction apparatuses and methods
US9076212B2 (en) 2006-05-19 2015-07-07 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US20150241171A1 (en) * 2014-02-26 2015-08-27 Supas Ltd Scope adjustment device
WO2015157102A3 (en) * 2014-04-08 2015-12-03 Eon Reality, Inc. Interactive virtual reality systems and methods
US20150377588A1 (en) * 2014-06-30 2015-12-31 Daniel James Periard Method and Apparatus for Aligning Laser to Optical System
US9261332B2 (en) 2013-05-09 2016-02-16 Shooting Simulator, Llc System and method for marksmanship training
US9267762B2 (en) 2013-05-09 2016-02-23 Shooting Simulator, Llc System and method for marksmanship training
US9305365B2 (en) 2013-01-24 2016-04-05 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
TWI555965B (en) * 2015-07-20 2016-11-01 信泰光學(深圳)有限公司 Tracking-monitoring method, tracking-monitoring device and operational method for tracking-monitoring device
US9542011B2 (en) 2014-04-08 2017-01-10 Eon Reality, Inc. Interactive virtual reality systems and methods
US20170052677A1 (en) * 2015-08-21 2017-02-23 Bubble Llc System and method for presenting an object
WO2017035033A1 (en) * 2015-08-21 2017-03-02 Bubble Llc System and method for presenting an object
US9606209B2 (en) 2011-08-26 2017-03-28 Kineticor, Inc. Methods, systems, and devices for intra-scan motion correction
US9684369B2 (en) 2014-04-08 2017-06-20 Eon Reality, Inc. Interactive virtual reality systems and methods
US9717461B2 (en) 2013-01-24 2017-08-01 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9734589B2 (en) 2014-07-23 2017-08-15 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9782141B2 (en) 2013-02-01 2017-10-10 Kineticor, Inc. Motion tracking system for real time adaptive motion compensation in biomedical imaging
US20180031353A1 (en) * 2012-10-16 2018-02-01 Nicholas Chris Skrepetos Systems, methods, and devices for electronically displaying individual shots from multiple shots on one physical target
US9943247B2 (en) 2015-07-28 2018-04-17 The University Of Hawai'i Systems, devices, and methods for detecting false movements for motion correction during a medical imaging scan
FR3059091A1 (en) * 2016-11-24 2018-05-25 Airbus Ds Sas SIMULATION SHOOTING GLASS
WO2018100487A1 (en) * 2016-12-01 2018-06-07 Sidelkovsky Felix Methods systems circuits components apparatus devices assemblies and computer executable code for aiming a firearm
US10004462B2 (en) 2014-03-24 2018-06-26 Kineticor, Inc. Systems, methods, and devices for removing prospective motion correction from medical imaging scans
US10030937B2 (en) 2013-05-09 2018-07-24 Shooting Simulator, Llc System and method for marksmanship training
US10041764B2 (en) 2016-01-09 2018-08-07 Han Shyone Ooi Sensor system and method for sensing motion
US20180372440A1 (en) * 2017-06-22 2018-12-27 Cubic Corporation Weapon barrel attachment for triggering instrumentation laser
US10234240B2 (en) 2013-05-09 2019-03-19 Shooting Simulator, Llc System and method for marksmanship training
US10274287B2 (en) 2013-05-09 2019-04-30 Shooting Simulator, Llc System and method for marksmanship training
US10309751B2 (en) 2016-04-28 2019-06-04 Cole Engineering Services, Inc. Small arms shooting simulation system
US10327708B2 (en) 2013-01-24 2019-06-25 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
WO2019144088A1 (en) * 2018-01-22 2019-07-25 Hvrt Corp. Systems and methods for shooting simulation and training
US10495416B2 (en) * 2013-01-10 2019-12-03 Brian Donald Wichner Methods and systems for determining a gunshot sequence or recoil dynamics of a gunshot for a firearm
US10527920B1 (en) * 2018-06-14 2020-01-07 Dhpc Technologies, Inc. System, method and device for a high fidelity electro-optical simulator
US10584940B2 (en) 2013-05-09 2020-03-10 Shooting Simulator, Llc System and method for marksmanship training
US10613426B1 (en) * 2018-06-14 2020-04-07 Dhpc Technologies, Inc. System, method and device for a long range, real size weapon systems plume simulator for testing optical detection devices in the field
US10716515B2 (en) 2015-11-23 2020-07-21 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US10739109B1 (en) 2016-10-28 2020-08-11 Selso Tello Firearm marksmanship system with chamber insert
US11150047B2 (en) 2019-07-25 2021-10-19 The Boeing Company Firearm incorporating recoil reduction
US11420248B2 (en) 2018-12-12 2022-08-23 The Boeing Company Impact-compensating bucking bar
CN115038928A (en) * 2020-02-03 2022-09-09 贝以系统哈格伦斯公司 Embedded target tracking training
US20230304763A1 (en) * 2017-01-27 2023-09-28 Armaments Research Company, Inc. Weapon usage monitoring system having performance metrics based on discharge event detection
US11774323B1 (en) 2021-03-25 2023-10-03 Dhpc Technologies, Inc. System and method for creating a collimated space for a high fidelity simulator

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020012898A1 (en) * 2000-01-13 2002-01-31 Motti Shechter Firearm simulation and gaming system and method for operatively interconnecting a firearm peripheral to a computer system
US6458034B1 (en) * 1999-08-27 2002-10-01 Namco Ltd. Game system and computer-usable information
US6582299B1 (en) * 1998-12-17 2003-06-24 Konami Corporation Target shooting video game device, and method of displaying result of target shooting video game
US20040031184A1 (en) * 2000-01-14 2004-02-19 Hope Richard W. Optical imaging device for firearm scope attachment
US6942486B2 (en) * 2003-08-01 2005-09-13 Matvey Lvovskiy Training simulator for sharp shooting
US20050233284A1 (en) * 2003-10-27 2005-10-20 Pando Traykov Optical sight system for use with weapon simulation system
US20090059364A1 (en) * 2006-11-02 2009-03-05 Sensics Inc. Systems and methods for electronic and virtual ocular devices
US20090155747A1 (en) * 2007-12-14 2009-06-18 Honeywell International Inc. Sniper Training System
US20120021385A1 (en) * 2006-11-24 2012-01-26 Trex Enterprises Corp. Celestial weapons orientation measuring system
US20120240444A1 (en) * 2011-03-25 2012-09-27 Robert Wayne Russell Camera Mount Apparatus and System for a Scope

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6582299B1 (en) * 1998-12-17 2003-06-24 Konami Corporation Target shooting video game device, and method of displaying result of target shooting video game
US6458034B1 (en) * 1999-08-27 2002-10-01 Namco Ltd. Game system and computer-usable information
US20020012898A1 (en) * 2000-01-13 2002-01-31 Motti Shechter Firearm simulation and gaming system and method for operatively interconnecting a firearm peripheral to a computer system
US20040031184A1 (en) * 2000-01-14 2004-02-19 Hope Richard W. Optical imaging device for firearm scope attachment
US6942486B2 (en) * 2003-08-01 2005-09-13 Matvey Lvovskiy Training simulator for sharp shooting
US20050233284A1 (en) * 2003-10-27 2005-10-20 Pando Traykov Optical sight system for use with weapon simulation system
US20090059364A1 (en) * 2006-11-02 2009-03-05 Sensics Inc. Systems and methods for electronic and virtual ocular devices
US20120021385A1 (en) * 2006-11-24 2012-01-26 Trex Enterprises Corp. Celestial weapons orientation measuring system
US20090155747A1 (en) * 2007-12-14 2009-06-18 Honeywell International Inc. Sniper Training System
US20120240444A1 (en) * 2011-03-25 2012-09-27 Robert Wayne Russell Camera Mount Apparatus and System for a Scope

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Anonymous. "Silent Scope: Operator's Manual", Konami. 1999. Retrieved from the internet 10 OCT 2012. Retrieved from *

Cited By (86)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
USRE44786E1 (en) 2005-12-22 2014-03-04 Force Science Institute Ltd. System and method for monitoring handling of a firearm or other trigger-based device
US9076212B2 (en) 2006-05-19 2015-07-07 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US9138175B2 (en) 2006-05-19 2015-09-22 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US10869611B2 (en) 2006-05-19 2020-12-22 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US9867549B2 (en) 2006-05-19 2018-01-16 The Queen's Medical Center Motion tracking system for real time adaptive imaging and spectroscopy
US20130040268A1 (en) * 2010-04-23 2013-02-14 Tacktech Pty (Ltd) Simulated shooting device and system
US20110315767A1 (en) * 2010-06-28 2011-12-29 Lowrance John L Automatically adjustable gun sight
US20120115110A1 (en) * 2010-11-10 2012-05-10 Christian Patrick Wade Method and tool for firearm sight training
US8684738B2 (en) * 2010-11-10 2014-04-01 The United States Of America, As Represented By The Secretary Of The Navy Method and tool for firearm sight training
US10663553B2 (en) 2011-08-26 2020-05-26 Kineticor, Inc. Methods, systems, and devices for intra-scan motion correction
US9606209B2 (en) 2011-08-26 2017-03-28 Kineticor, Inc. Methods, systems, and devices for intra-scan motion correction
WO2013165497A2 (en) * 2012-03-05 2013-11-07 Millett James A D-scope aiming device
US8807430B2 (en) * 2012-03-05 2014-08-19 James Allen Millett Dscope aiming device
WO2013165497A3 (en) * 2012-03-05 2014-01-16 Millett James A D-scope aiming device
US8881981B2 (en) 2012-03-05 2014-11-11 James A. Millett Digital targeting scope apparatus
CN104246420A (en) * 2012-03-05 2014-12-24 詹姆斯·A·米利特 D-scope aiming device
US20130228618A1 (en) * 2012-03-05 2013-09-05 James Allen Millett Dscope aiming device
US9140521B2 (en) 2012-03-05 2015-09-22 James A. Millett D-scope aiming device
US8857714B2 (en) * 2012-03-15 2014-10-14 Flir Systems, Inc. Ballistic sight system
US8950102B1 (en) 2012-04-13 2015-02-10 The Board Of Trustees Of The University Of Alabama Scope correction apparatuses and methods
US20130288205A1 (en) * 2012-04-30 2013-10-31 Trackingpoint, Inc. Rifle Scope and Method of Providing Embedded Training
US10480903B2 (en) * 2012-04-30 2019-11-19 Trackingpoint, Inc. Rifle scope and method of providing embedded training
US20130344461A1 (en) * 2012-06-26 2013-12-26 Selso Tello Universal Firearm Marksmanship System
US10145653B2 (en) 2012-06-26 2018-12-04 Selso Tello Universal firearm marksmanship system
US10900749B2 (en) 2012-06-26 2021-01-26 Selso Tello Universal firearm marksmanship system
US11506468B2 (en) 2012-06-26 2022-11-22 Selso Tello Universal firearm marksmanship system
US9335125B2 (en) * 2012-06-26 2016-05-10 Selso Tello Universal firearm marksmanship system
US20180031353A1 (en) * 2012-10-16 2018-02-01 Nicholas Chris Skrepetos Systems, methods, and devices for electronically displaying individual shots from multiple shots on one physical target
US10247517B2 (en) * 2012-10-16 2019-04-02 Nicholas Chris Skrepetos Systems, methods, and devices for electronically displaying individual shots from multiple shots on one physical target
US20140182187A1 (en) * 2012-12-31 2014-07-03 Trackingpoint, Inc. Software-Extensible Gun Scope and Method
US10495416B2 (en) * 2013-01-10 2019-12-03 Brian Donald Wichner Methods and systems for determining a gunshot sequence or recoil dynamics of a gunshot for a firearm
US10327708B2 (en) 2013-01-24 2019-06-25 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9607377B2 (en) 2013-01-24 2017-03-28 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US10339654B2 (en) 2013-01-24 2019-07-02 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US9717461B2 (en) 2013-01-24 2017-08-01 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9305365B2 (en) 2013-01-24 2016-04-05 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US9779502B1 (en) 2013-01-24 2017-10-03 Kineticor, Inc. Systems, devices, and methods for tracking moving targets
US10653381B2 (en) 2013-02-01 2020-05-19 Kineticor, Inc. Motion tracking system for real time adaptive motion compensation in biomedical imaging
US9782141B2 (en) 2013-02-01 2017-10-10 Kineticor, Inc. Motion tracking system for real time adaptive motion compensation in biomedical imaging
US9033711B2 (en) * 2013-03-15 2015-05-19 Kenneth W Guenther Interactive system and method for shooting and target tracking for self-improvement and training
WO2014145079A1 (en) * 2013-03-15 2014-09-18 Jacob Stewart Schmehl Firearm accessory
US20140272807A1 (en) * 2013-03-15 2014-09-18 Kenneth W. Guenther Interactive system and method for shooting and target tracking for self-improvement and training
US10274287B2 (en) 2013-05-09 2019-04-30 Shooting Simulator, Llc System and method for marksmanship training
US10030937B2 (en) 2013-05-09 2018-07-24 Shooting Simulator, Llc System and method for marksmanship training
US10584940B2 (en) 2013-05-09 2020-03-10 Shooting Simulator, Llc System and method for marksmanship training
US9267762B2 (en) 2013-05-09 2016-02-23 Shooting Simulator, Llc System and method for marksmanship training
US9261332B2 (en) 2013-05-09 2016-02-16 Shooting Simulator, Llc System and method for marksmanship training
US10234240B2 (en) 2013-05-09 2019-03-19 Shooting Simulator, Llc System and method for marksmanship training
US20150241171A1 (en) * 2014-02-26 2015-08-27 Supas Ltd Scope adjustment device
US9651338B2 (en) * 2014-02-26 2017-05-16 Supas Ltd Scope adjustment device
US10004462B2 (en) 2014-03-24 2018-06-26 Kineticor, Inc. Systems, methods, and devices for removing prospective motion correction from medical imaging scans
US9684369B2 (en) 2014-04-08 2017-06-20 Eon Reality, Inc. Interactive virtual reality systems and methods
WO2015157102A3 (en) * 2014-04-08 2015-12-03 Eon Reality, Inc. Interactive virtual reality systems and methods
CN106061571A (en) * 2014-04-08 2016-10-26 Eon现实公司 Interactive virtual reality systems and methods
US9542011B2 (en) 2014-04-08 2017-01-10 Eon Reality, Inc. Interactive virtual reality systems and methods
US20150377588A1 (en) * 2014-06-30 2015-12-31 Daniel James Periard Method and Apparatus for Aligning Laser to Optical System
US11100636B2 (en) 2014-07-23 2021-08-24 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US9734589B2 (en) 2014-07-23 2017-08-15 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US10438349B2 (en) 2014-07-23 2019-10-08 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
TWI555965B (en) * 2015-07-20 2016-11-01 信泰光學(深圳)有限公司 Tracking-monitoring method, tracking-monitoring device and operational method for tracking-monitoring device
US10660541B2 (en) 2015-07-28 2020-05-26 The University Of Hawai'i Systems, devices, and methods for detecting false movements for motion correction during a medical imaging scan
US9943247B2 (en) 2015-07-28 2018-04-17 The University Of Hawai'i Systems, devices, and methods for detecting false movements for motion correction during a medical imaging scan
WO2017035033A1 (en) * 2015-08-21 2017-03-02 Bubble Llc System and method for presenting an object
US10402209B2 (en) * 2015-08-21 2019-09-03 Bubble Llc System and method for presenting an object
US20170052677A1 (en) * 2015-08-21 2017-02-23 Bubble Llc System and method for presenting an object
US10716515B2 (en) 2015-11-23 2020-07-21 Kineticor, Inc. Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan
US10041764B2 (en) 2016-01-09 2018-08-07 Han Shyone Ooi Sensor system and method for sensing motion
US10309751B2 (en) 2016-04-28 2019-06-04 Cole Engineering Services, Inc. Small arms shooting simulation system
US11002512B2 (en) 2016-10-28 2021-05-11 Selso Tello Firearm marksmanship system with chamber insert
US10739109B1 (en) 2016-10-28 2020-08-11 Selso Tello Firearm marksmanship system with chamber insert
WO2018096023A1 (en) * 2016-11-24 2018-05-31 Airbus Ds Sas Firing simulation scope
FR3059091A1 (en) * 2016-11-24 2018-05-25 Airbus Ds Sas SIMULATION SHOOTING GLASS
US11268790B2 (en) * 2016-11-24 2022-03-08 Airbus Defence And Space S.A.S. Firing-simulation scope
WO2018100487A1 (en) * 2016-12-01 2018-06-07 Sidelkovsky Felix Methods systems circuits components apparatus devices assemblies and computer executable code for aiming a firearm
US20230304763A1 (en) * 2017-01-27 2023-09-28 Armaments Research Company, Inc. Weapon usage monitoring system having performance metrics based on discharge event detection
US20180372440A1 (en) * 2017-06-22 2018-12-27 Cubic Corporation Weapon barrel attachment for triggering instrumentation laser
US10788290B2 (en) * 2018-01-22 2020-09-29 Hvrt Corp. Systems and methods for shooting simulation and training
WO2019144088A1 (en) * 2018-01-22 2019-07-25 Hvrt Corp. Systems and methods for shooting simulation and training
US20190226808A1 (en) * 2018-01-22 2019-07-25 Hvrt Corp. Systems and methods for shooting simulation and training
US11754372B2 (en) 2018-01-22 2023-09-12 Hvrt Corp. Systems and methods for shooting simulation and training
US10613426B1 (en) * 2018-06-14 2020-04-07 Dhpc Technologies, Inc. System, method and device for a long range, real size weapon systems plume simulator for testing optical detection devices in the field
US10527920B1 (en) * 2018-06-14 2020-01-07 Dhpc Technologies, Inc. System, method and device for a high fidelity electro-optical simulator
US11420248B2 (en) 2018-12-12 2022-08-23 The Boeing Company Impact-compensating bucking bar
US11150047B2 (en) 2019-07-25 2021-10-19 The Boeing Company Firearm incorporating recoil reduction
CN115038928A (en) * 2020-02-03 2022-09-09 贝以系统哈格伦斯公司 Embedded target tracking training
US11774323B1 (en) 2021-03-25 2023-10-03 Dhpc Technologies, Inc. System and method for creating a collimated space for a high fidelity simulator

Similar Documents

Publication Publication Date Title
US20110207089A1 (en) Firearm training systems and methods of using the same
US11506468B2 (en) Universal firearm marksmanship system
US10234240B2 (en) System and method for marksmanship training
US10584940B2 (en) System and method for marksmanship training
US10274287B2 (en) System and method for marksmanship training
CN101869765B (en) Shooting training systems and methods using an embedded photo sensing panel
US11015902B2 (en) System and method for marksmanship training
US8794967B2 (en) Firearm training system
TWI485630B (en) Sights, operational methods thereof, and computer program products thereof
US20070254266A1 (en) Marksmanship training device
US10480903B2 (en) Rifle scope and method of providing embedded training
US20120183931A1 (en) Hit detection in direct-fire or small-arms simulators
US6973865B1 (en) Dynamic pointing accuracy evaluation system and method used with a gun that fires a projectile under control of an automated fire control system
WO2021048307A1 (en) Imaging system for firearm
CN108398053A (en) Bearing calibration, device, computer equipment and the firearms of dynamic collimation
TWI642893B (en) Target acquisition device and system thereof
KR20130006113A (en) Apparatus for fire training simulation system
US20210048276A1 (en) Probabilistic low-power position and orientation
CN110631411A (en) Virtual shooting training control method and system
CA3130642A1 (en) Device and method for shot analysis
CN215064086U (en) Shooting range system
US20210164760A1 (en) Integrated shooting simulation system using fisheye lens camera
TWI436026B (en) Shooting training simulator aiming correction method
WO2023042195A1 (en) Smart aiming device with built-in training system for marksmanship and firearm operation
UA149085U (en) COMPLEX OF AUDIOVISUAL CORRECTION OF AIMING WHEN PERFORMING EXERCISES WITH SMALL WEAPONS

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION