USRE44225E1 - Abnormality detection and surveillance system - Google Patents

Abnormality detection and surveillance system Download PDF

Info

Publication number
USRE44225E1
USRE44225E1 US13/230,490 US201113230490A USRE44225E US RE44225 E1 USRE44225 E1 US RE44225E1 US 201113230490 A US201113230490 A US 201113230490A US RE44225 E USRE44225 E US RE44225E
Authority
US
United States
Prior art keywords
individual
movements
movement
criminal intent
criminal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US13/230,490
Inventor
David G. Aviv
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Prophet Productions LLC
Original Assignee
Prophet Productions LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=25409505&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=USRE44225(E1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Priority claimed from US08/367,712 external-priority patent/US5666157A/en
Application filed by Prophet Productions LLC filed Critical Prophet Productions LLC
Priority to US13/230,490 priority Critical patent/USRE44225E1/en
Application granted granted Critical
Publication of USRE44225E1 publication Critical patent/USRE44225E1/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19602Image analysis to detect motion of the intruder, e.g. by frame subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19602Image analysis to detect motion of the intruder, e.g. by frame subtraction
    • G08B13/19608Tracking movement of a target, e.g. by detecting an object predefined as a target, using target direction and or velocity to predict its new position
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19602Image analysis to detect motion of the intruder, e.g. by frame subtraction
    • G08B13/19613Recognition of a predetermined image pattern or behaviour pattern indicating theft or intrusion
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19602Image analysis to detect motion of the intruder, e.g. by frame subtraction
    • G08B13/19613Recognition of a predetermined image pattern or behaviour pattern indicating theft or intrusion
    • G08B13/19615Recognition of a predetermined image pattern or behaviour pattern indicating theft or intrusion wherein said pattern is defined by the user
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19639Details of the system layout
    • G08B13/19641Multiple cameras having overlapping views on a single scene
    • G08B13/19643Multiple cameras having overlapping views on a single scene wherein the cameras play different roles, e.g. different resolution, different camera type, master-slave camera
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19665Details related to the storage of video surveillance data
    • G08B13/19676Temporary storage, e.g. cyclic memory, buffer storage on pre-alarm
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • G08B13/19697Arrangements wherein non-video detectors generate an alarm themselves
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/188Capturing isolated or intermittent images triggered by the occurrence of a predetermined event, e.g. an object reaching a predetermined position
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S706/00Data processing: artificial intelligence
    • Y10S706/902Application using ai with detail of the ai system
    • Y10S706/933Law, law enforcement, or government

Definitions

  • This invention generally relates to surveillance systems, and more particularly, to trainable surveillance systems which detect and respond to specific abnormal video and audio input signals.
  • U.S. Pat. No. 4,737,847 issued to Araki et al. discloses an improved abnormality surveillance system wherein motion sensors are positioned within a protected area to first determine the presence of an object of interest, such as an intruder.
  • zones having prescribed “warning levels” are defined within the protected area. Depending on which of these zones an object or person is detected in, moves to, and the length of time the detected object or person remains in a particular zone determines whether the object or person entering the zone should be considered an abnormal event or a threat.
  • a surveillance system having at least one primary video camera for translating real images of a zone into electronic video signals at a first level of resolution
  • a method for determining criminal activity by an individual within a field of view of a video camera comprising:
  • FIG. 1 is a schematic block diagram of the video, analysis, control, alarm and recording subsystems embodying this invention
  • FIG. 2A illustrates a frame K of a video camera's output of a particular environment, according to the invention, showing four representative objects (people) A, B, C, and D, wherein objects A, B and D are moving in a direction indicated with arrows, and object C is not moving;
  • FIG. 2B illustrates a frame K+5 of the video camera's output, according to the invention, showing objects A, B, and D are stationary, and object C is moving;
  • FIG. 2C illustrates a frame K+10 of the video camera's output, according to the invention, showing the current location of objects A, B, C, D, and E;
  • FIG. 2D illustrates a frame K+11 of the video camera's output, according to the invention, showing object B next to object C, and object E moving to the right;
  • FIG. 2E illustrates a frame K+12 of the video camera's output, according to the invention, showing a potential crime taking place between objects B and C;
  • FIG. 2F illustrates a frame K+13 of the video camera's output, according to the invention, showing objects B and C interacting;
  • FIG. 2G illustrates a frame K+15 of the video camera's output, according to the invention, showing object C moving to the right and object B following;
  • FIG. 2H illustrates a frame K+16 of the video camera's output, according to the invention, showing object C moving away from a stationary object B;
  • FIG. 2I illustrates a frame K+17 of the video camera's output, according to the invention, showing object B moving towards object C.
  • FIG. 3A illustrates a frame of a video camera's output, according to the invention, showing a “two on one” interaction of objects (people) A, B, and C;
  • FIG. 3B illustrates a later frame of the video camera's output of FIG. 3A , according to the invention, showing objects A and C moving towards object B;
  • FIG. 3C illustrates a later frame of the video camera's output of FIG. 3B , according to the invention, showing objects A and C moving in close proximity to object B;
  • FIG. 3D illustrates a later frame of the video camera's output of FIG. 3C , according to the invention, showing objects A and C quickly moving away from object B.
  • FIG. 4 is a schematic block diagram of a conventional word recognition system
  • FIG. 5 is a schematic block diagram of a video and verbal recognition system, according to the invention.
  • picture input means 10 which may be any conventional electronic picture pickup device operational within the infrared or visual spectrum (or both) including a vidicon and a CCD/TV camera (including the wireless type).
  • picture input means 10 there is the deployment of a high rate camera/recorder (similar to those made by NAC Visual Systems of Woodland Hills, Calif., SONY and others).
  • a high rate camera/recorder similar to those made by NAC Visual Systems of Woodland Hills, Calif., SONY and others.
  • Such high rate camera/recorder systems are able to detect and record very rapid movements of body parts that are commonly indicative of a criminal intent. Such fast movements might not be resolved with a more standard 30 frames per second camera. However, most movements will be resolved with a standard 30 frames per second camera.
  • This picture means may also be triggered by an alert signal from the processor of the low resolution camera or, as before, from the audio/word recognition processor when sensing a suspicious event.
  • the primary picture input means 10 is preferably a low cost video camera wherein high resolution is not necessary and due to the relative expense will most likely provide only moderate resolution.
  • the preferred CCD/TV camera is about 11 ⁇ 2 inches in length and about 1 inch in diameter, weighing about 3 ounces, and for particular deployment, a zoom lens attachment may be used).
  • This device will be operating continuously and will translate the field of view (“real”) images within a first observation area into conventional video electronic signals.
  • picture input means 10 In another embodiment of picture input means 10 , a high rate camera/recorder, (similar to those made by NAC Visual Systems of Woodland Hills, Calif., SONY and others) is used, which would then enable the detection of even the very rapid movement of body parts that are indicative of criminal intent, and their recording.
  • the more commonly used camera operates at 30 frames per second will be able to resolve essentially all quick body movements.
  • the picture input means may also be activated by an “alert” signal from the processor of the low resolution camera or from the audio/word recognition processor when sensing a suspicious event.
  • the picture input means for any embodiment contains a preprocessor which normalizes a wide range of illumination levels, especially for outside observation.
  • the preprocessor to emulates a vertebrate's retina, which has an efficient and accurate normalization process.
  • One such preprocessor (VLSI retina chip) is fabricated by the Carver Meade Laboratory of the California Institute of Technology in Pasadena, Calif. Use of this particular preprocessor chip will increase the automated vision capability of this invention whenever variation of light intensity and light reflection may otherwise weaken the picture resolution.
  • the signals from the picture input means 10 are converted into digitized signals and then sent to the picture processing means 12 .
  • the processor controlling each group of cameras will be governed by an artificial intelligence system, based on dynamic pattern recognition principles, as further described below.
  • the picture processing means 12 includes an image raster analyzer which effectively segments each image to isolate each pair of people.
  • the image raster analyzer subsystem of picture processing means 12 segments each sampled image to identify and isolate each pair of objects (or people), and each “two on one” group of 3 people separately.
  • the “2 on 1” represents a common mugging situation in which two individuals approach a victim: one from in front of the victim and the other from behind.
  • the forward mugger tells the potential victim that if he does not give up his money, (or watch, ring, etc.) the second mugger will shoot him, stab or otherwise harm him.
  • the group of three people will thus be considered a potential crime in progress and will therefore be segmented and analyzed in picture processing means.
  • An additional embodiment of the picture means 1 is the inclusion of an optics system known as the zoom lens system.
  • the essentials of the zoom lens subsystem are described in three papers written by L. Motz and L. Bergstein, in an article titled “Zoom Lens Systems” in the Journal of Optical Society of America, Vol. 52, April, 1992. This article is hereby incorporated by reference.
  • the essence of the zoom system is to vary the focal length such that an object being observed will be focused and magnified at its image plane.
  • FOV field-of-view
  • the lens which moves to focus the object onto the camera's image plane.
  • An error which is used to correct the focus, by the image planes's is generated by CCD array into 2 halves and measuring the difference segmenting in each until the object is at the center. Dividing the CCD array into more than 2 segments, say 4 quadrants is a way to achieve automatic centering, as is the case with mono-pulse radar. Regardless of the number of segments, the error signal is used to generate the desired tracking of the object.
  • the zoom with input from the segmentation subsystem of the picture analysis means 12 will focus on the object closest to the right hand side of the image plane, and then proceed to move the focus to the left, focusing on the next object and on the next sequentially.
  • the automatic zoom can more naturally choose to home-in on the person with the brightest emission or reflection, and then proceed to the next brightness and so forth. This would be a form of an intensity/time selection multiplex zoom system.
  • the relative positioning of the input camera with respect to the area under surveillance will effect the accuracy by which the image raster analyzer segments each image.
  • the height of the input camera is preferably sufficient to minimize occlusion between the input camera and the movement of the individuals under surveillance.
  • Each image frame segment, once digitized, is stored in a frame by frame memory storage of section 12 .
  • Each frame from the camera input 10 is subtracted from a previous frame already stored in memory 12 using any conventional differencing process.
  • the differencing process involving multiple differencing steps takes place in the differencing section 12 .
  • the resulting difference signal (outputted from the differencing sub-section 12 ) of each image indicates all the changes that have occurred from one frame to the next. These changes include any movements of the individuals located within the segment and any movements of their limbs, e.g., arms.
  • a collection of differencing signals for each moved object of subsequent sampled frames of images allows a determination of the type, speed and direction (vector) of each motion involved and also processing which will extract acceleration, i.e., note of change of velocity: and change in acceleration with respect to time (called “jerkiness”) and will when correlating with stored signatures of known physical criminal acts.
  • subsequent differencing signals may reveal that an individual's arm is moving to a high position, such as the upper limit of that arm's motion, i.e., above his head) at a fast speed. This particular movement could be perceived, as described below, as a hostile movement with a possible criminal intent requiring the expert analysis of security personnel.
  • the intersection of two tracks indicates the intersection of two moved objects.
  • the intersecting objects in this case, could be merely the two hands of two people greeting each other, or depending on other characteristics, as described below, the intersecting objects could be interpreted as a fist of an assailant contacting the face of a victim in a less friendly greeting.
  • the intersection of two tracks immediately requires further analysis and/or the summoning of security personnel. But the generation of an alarm, light and sound devices located, for example, on a monitor will turn a guard's attention only to that monitor, hence the labor savings.
  • friendly interactions between individuals is a much slower physical process than is a physical assault vis-a-vis body parts of the individuals involved. Hence, friendly interactions may be easily distinguished from hostile physical acts using current low pass and high pass filters, and current pattern recognition techniques based on experimental reference data.
  • optical flow computation A commercially available software tool may enhance object-movement analysis between frames (called optical flow computation).
  • optical flow computation specific (usually bright) reflective elements, called farkles, emitted from the clothing and/or the body parts of an individual of one frame are subtracted from a previous frame.
  • the bright portions will inherently provide sharper detail and therefore will yield more accurate data regarding the velocities of the relative moving objects.
  • Additional computation as described below, will provide data regarding the acceleration and even change in acceleration or “jerkiness” of each moving part sampled.
  • the physical motions of the individuals involved in an interaction will be detected by first determining the edges of the of each person imaged. And the movements of the body parts will then be observed by noting the movements of the edges of the body parts of the (2 or 3) individuals involved in the interaction.
  • the differencing process will enable the determination of the velocity and acceleration and rate of acceleration of those body parts.
  • the now processed signal is sent to comparison means 14 which compares selected frames of the video signals from the picture input means 10 with “signature” video signals stored in memory 16 .
  • the signature signals are representative of various positions and movements of the body ports of an individual having various levels of criminal intent. The method for obtaining the data base of these signature video signals in accordance with another aspect of the invention is described in greater detail below.
  • an output “alert” signal is sent from the comparison means 14 to a controller 18 .
  • the controller 18 controls the operation of a secondary, high resolution picture input means (video camera) 20 and a conventional monitor 22 and video recorder 24 .
  • the field of view of the secondary camera 20 is preferably at most, the same as the field of view of the primary camera 10 , surveying a second observation area.
  • the recorder 24 may be located at the site and/or at both a law enforcement facility (not shown) and simultaneously at a Court office or legal facility to prevent loss of incriminating information due to tampering.
  • the purpose of the secondary camera 20 is to provide a detailed video signal of the individual having assumed criminal intent and also to improve false positive and false negative performance. This information is recorded by the video recorder 24 and displayed on a monitor 22 . An alarm bell or light (not shown) or both may be provided and activated by an output signal from the controller 20 to summon a supervisor to immediately view the pertinent video images showing the apparent crime in progress and access its accuracy.
  • a VCR 26 is operating continuously (using a 6 hour loop-tape, for example).
  • the VCR 26 is being controlled by the VCR controller 28 .
  • All the “real-time” images directly from the picture input means 10 are immediately recorded and stored for at least 6 hours, for example.
  • a signal from the controller 18 is sent to the VCR controller 28 changing the mode of recording from tape looping mode to non-looping mode.
  • the tape will not re-loop and will therefore retain the perhaps vital recorded video information of the surveyed site, including the crime itself, and the events leading up to the crime.
  • the video signal may also be transmitted to a VCR located elsewhere; for example, at a law enforcement facility and, simultaneously to other secure locations of the Court and its associated offices.
  • each sampled frame of video is “segmented” into parts relating to the objects detected therein.
  • the video signal derived from the vidicon or CCD/TV camera is analyzed by an image raster analyzer. Although this process causes slight signal delays, it is accomplished nearly in real time.
  • a high resolution camera may not be required or otherwise used.
  • the resolution provided by a relatively simple and low cost camera may be sufficient.
  • the length of frame intervals between analyzed frames may vary. For example, in a high risk area, every frame from the CCD/TV camera may be analyzed continuously to ensure that the maximum amount of information is recorded prior to and during a crime. In a low risk area, it may be preferred to sample perhaps every 10 frames from each camera, sequentially.
  • the system would activate an alert mode wherein the system becomes “concerned and curious” in the suspicious actions and the sampling rate is increased to perhaps every 5 frames or even every frame.
  • an alert mode wherein the system becomes “concerned and curious” in the suspicious actions and the sampling rate is increased to perhaps every 5 frames or even every frame.
  • the entire system may be activated wherein both audio and video system begin to sample the environment for sufficient information to determine the intent of the actions.
  • FIG. 2 several frames of a particular camera output are shown to illustrate the segmentation process performed in accordance with the invention.
  • the system begins to sample at frame K and determines that there are four objects (previously determined to be people, as described below), A-D located within a particular zone being policed. Since nothing unusual is determined from the initial analysis, the system does not warrant an “alert” status. People A, B, and D are moving according to normal, non-criminal intent, as could be observed.
  • a crime likelihood is indicated when frames K+10 through K+13 are analyzed by the differencing process. And if the movement of the body parts indicate velocity, acceleration and “jerkiness” that compare positively with the stored digital signals depicting movements of known criminal physical assaults, it is likely that a crime is in progress here.
  • An alarm is generated the instant any of the above conditions is established.
  • This alarm condition will result in sending in police or Guards to the crime site, activating the high resolution CCD/TV camera to record the face of the person committing the assault, a loud speaker being activated automatically, playing a recorded announcement warning the perpetrator the seriousness of his actions now being undertaken and demanding that he cease the criminal act. After dark a strong light will be turned on automatically.
  • the automated responses will be actuated the instant an alarm condition is adjudicated by the processor.
  • an alarm signal is sent to the police station and the same video signal of the event, is transmitted to a court appointed data collection office, to the Public Defender's office and the District Attorney's Office.
  • the present invention could be easily implemented at various sites to create effective “Crime Free” zones.
  • the above described Abnormality Detection System includes an RF-ID (Radio Frequency Identification) tag, to assist in the detection and tracking of individuals within the field of view of a camera.
  • RF-ID Radio Frequency Identification
  • tags are worn by authorized individuals.
  • the tags response when queried by the RF Interrogator.
  • the response signal of the tags propagation pattern which is adequately registered with the video sensor.
  • the “Tags” are sensed in video are assumed friendly and authorized. This information will simplify the segmentation process.
  • each RF-ID card will be turned ON, when a positive response to an interrogation signal is established.
  • the light will appear on the computer generated grid (also on the screen of the monitor) and the intersection of tracks clearly indicated, followed by their physical interaction. But also noted will be the intersection between the tagged and the untagged individuals. In all of such cases, the segmentation process will be simpler.
  • the applications of the present invention include stationary facilities: banks and ATMs, hotels, private residence halls and dormitories, high rise and low rise office and residential buildings, public and private schools from kindergarten through high-school, colleges and universities, hospitals, sidewalks, street crossing, parks, containers and container loading areas, shipping piers, train stations, truck loading stations, airport passenger and freight facilities, bus stations, subway stations, move houses, theaters, concert halls and arenas, sport arenas, libraries, churches, museums, stores, shopping malls, restaurants, convenience stores, bars, coffee shops, gasoline stations, highway rest stops, tunnels, bridges, gateways, sections of highways, toll booths, warehouses, and depots, factories and assembly rooms, law enforcement facilities including jails.
  • Further applications of the invention include areas of moving platforms: automobiles, trucks, buses, subway cars, train cars, freight and passenger, boats and ships (passenger and freight, tankers, service vehicles, construction vehicles, on and off-road, containers and their carriers, and airplanes.
  • military applications that will include but will not be limited to assorted military ground, sea, and air mobile vehicles and assorted military ground, sea, and air mobile vehicles and platforms as well as stationary facilities where the protection of low, medium, and high value targets are necessary; such targets are common in the military but have equivalents in the civilian areas wherein this invention will serve both sectors.
  • a tiny CCD/TV camera connected surreptitiously at the ceiling of the car, or in the rear-view mirror, through a pin hole lens and focused at the driver's seat will be connected to the video processor to record the face of the drive.
  • the camera is triggered by the automatic word recognition processor that will identify the well known expressions commonly used by the car-jacker.
  • the video picture will be recorded and then transmitted via cellular phone in the car. Without a phone, the short video recording of the face of the car-jacker will be held until the car is found by the police, but now with the evidence (the picture of the car-jacker) in hand.
  • the security personnel manning the monitors are alerted only to video images which show suspicious actions (criminal activities) within a prescribed observation zone.
  • the security personnel are therefore used to access the accuracy of the crime and determine the necessary actions for an appropriate response.
  • computers to effectively filter out all normal and noncriminal video signals from observation areas, fewer security personnel are required to survey and “secure” a greater overall area (including a greater number of observation areas, i.e., cameras).
  • a battery operated portable version of the video system would automatically identify known objects in its field of view and a speech synthesizer would “say” the object. For example, “chair”, “table”, etc. would indicate the presence of a chair and a table.
  • At least two and perhaps three cameras are used simultaneously to cover the area. Should one camera sense a first level of criminal action, the other two could be manipulated to provide a three dimensional perspective coverage of the action.
  • the three dimensional image of a physical interaction in the policed area would allow observation of a greater number of details associated with the steps: accost, threat, assault, response and post response.
  • the conversion from the two dimensional image to the three dimensional image is known as “random transform”.
  • both video and acoustic information is sampled and analyzed.
  • the acoustic information is sampled and analyzed in a similar manner to the sampling and analyzing of the above-described video information.
  • the audio information is sampled and analyzed in a manner shown in FIG. 4 , and is based on prior art. (references 6 and 7).
  • ASR Automatic Speech Recognition
  • a conventional automatic word recognition system including an input microphone system 40 , an analysis subsystem 42 , a template subsystem 44 , a pattern comparator 46 , and a post-processor and decision logic subsystem 48 .
  • the acoustic/audio policing system will begin sampling all (or a selected portion) of nearby acoustic signals.
  • the acoustic signals will include voices and background noise.
  • the background noise signals are generally known and predictable, and may therefore be easily filtered out using conventional filtering techniques.
  • the expected noise signals are unfamiliar speech, automotive related sounds, honking, sirens, the sound of wind and/or rain.
  • the microphone input system 40 pick-up the acoustic signals and immediately filter out the predictable background noise signals and amplify the remaining recognizable acoustic signals.
  • the filtered acoustic signals are analyzed in the analysis subsystem 42 which processes the signals by means of digital and spectral analysis techniques.
  • the output of the analysis subsystem is compared in the pattern comparater subsystem 46 with selected predetermined words stored in memory in 44 .
  • the post processing and decision logic subsystem 48 generates an alarm signal, as described below.
  • the templates 44 include perhaps about 100 brief and easily recognizable terse expressions, some of which are single words, and are commonly used by those intent on a criminal act.
  • Some examples of commonly used word phrases spoken by a criminal to a victim prior to a mugging include: “Give me your money”, “This is a stick-up”, “Give me your wallet and you won't get hurt” . . . etc.
  • commonly used replies from a typical victim during such a mugging may also be stored as template words, such as “help”, and certain sounds such as shrieks, screams and groans, etc.
  • the specific word templates, from which inputted acoustic sounds are compared with, must be chosen carefully, taking into account the particular accents and slang of the language spoken in the region of concern (e.g., the southern cities of the U.S. will require a different template 44 than the one used for a recognition system in the New York City region of the U.S.).
  • the output of the word recognition system shown in FIG. 4 is used as a trigger signal to activate a sound recorder, or a camera used elsewhere in the invention, as described below.
  • the preferred microphone used in the microphone input subsystem 40 is a shotgun microphone, such as those commercially available from the Sennheiser Company of Frankfurt, Germany. These microphone have a super-cardioid propagation pattern. However, the gain of the pattern may be too small for high traffic areas and may therefore require more than one microphone in an array configuration to adequately focus and track in these areas. The propagation pattern of the microphone system enables better focusing on a moving sound source (e.g., a person walking and talking).
  • a conventional directional microphone may also be used in place of a shot-gun type microphone, such as those made by the Sony Corporation of Tokyo, Japan. Such directional microphones will achieve similar gain to the shot-gun type microphones, but with a smaller physical structure.
  • a feedback loop circuit (not specifically shown) originating in the post processing subsystem 48 will direct the microphone system to track a particular dynamic source of sound within the area surveyed by video cameras.
  • An override signal from the video portion of the present invention will activate and direct the microphone system towards the direction of the field of view of the camera.
  • the video system will control the audio recording system towards the scene of interest.
  • the audio system will direct appropriate video cameras to visually cover and record the apparent source of the sound.
  • HMM hidden Markov model
  • ANN artificial neural network
  • the HMM system uses probability statistics to predict a particular spoken word following recognition of a primary word unit, syllable or phoneme. For example, as the word “money” is inputted into an HMM word recognition system, the first recognized portion of the word is “mon . . . ”. The HMM system immediately recognizes this word stem and determines that the spoken word could be “MONDAY”, “MONopoly”, or “MONey”, etc. The resulting list of potential words is considerably shorter than the entire list of all spoken words of the English language. Therefore, the HMM system employed with the present invention allows both the audio and video systems to operate quickly and use HMM probability statistics to predict future movements or words based on an early recognition of initial movements and word stems.
  • the HMM system may be equally employed in the video recognition system. For example, if a person's arm quickly moves above his head, the HMM system may determine that there is a high probability that the arm will quickly come down, perhaps indicating a criminal intent.
  • the above-described system actively compares input data signals from a video camera, for example, with known reference data of specific body movements stored in memory.
  • reference data or ground truth data
  • This reference data describes threats, actual criminal physical acts, verbal threats and verbal assaults, and also friendly physical acts and friendly words, and neutral interactions between interacting people.
  • the reference data may be obtained using any of at least the following described three methods including a) attaching accelerometers at predetermined points (for example arm and leg joints, hips, and the forehead) of actors; b) using a computer to derive 3-D models of people (stored in the computer's memory as pixel data) and analyze the body part movements of the people; and c) scanning (or otherwise downloading) video data from movie and TV clips of various physical and verbal interactions into a computer to analyze specific movements and sounds.
  • the preferred method for obtaining reference data is includes attaching accelerometers to actors while performing various actions or “events” of interest: abnormal (e.g., criminal or generally quick, violent movements), normal (e.g., shaking hands, slow and smooth movements), and neutral behavior (e.g., walking).
  • abnormal e.g., criminal or generally quick, violent movements
  • normal e.g., shaking hands, slow and smooth movements
  • neutral behavior e.g., walking
  • the people located within the environment are provided personal ID cards that include an electronic radio frequency (rf) transmitter.
  • the transmitter of each radio-frequency identification card (RFID) transmits an rf signal that identifies the person carrying the card.
  • Receivers located in the area of a surveillance camera can receive the identification information and use it to help identify the different people located within the field of the near by surveillance camera (or microphone, in the case of audio analysis).
  • people may be issued an RFID card prior to entering a particular area, such as a U.S. Tennis Open event.
  • a clearance check would be made for each person prior to them receiving such a card.
  • surveillance cameras would associate card-holders as less likely to cause trouble and would be suspicious of anyone within the field of the camera's view not being identified by an RFID card.
  • the basic configuration of the invention uses video and audio sensors (such as, respectively, a camera and a microphone), and potentially other active and passive sensing and processing devices and systems (including the use of radar and ladar and other devices that operate in all areas of the electromagnetic spectrum) to detect threats and actual criminal acts occurring with a field of view of a camera (a video sensor).
  • video and audio sensors such as, respectively, a camera and a microphone
  • other active and passive sensing and processing devices and systems including the use of radar and ladar and other devices that operate in all areas of the electromagnetic spectrum
  • the system described above, and according to the invention initially requires the collection of “reference values” which correspond to specific known acts of threat, actual assault (both physical and verbal), and other physical and verbal interactions that are considered friendly or neutral.
  • Video components of recorded “reference data” is stored in a physical movement dictionary (or data base), while audio components of such reference data is stored in a verbal utterance dictionary (or data base).
  • real time (or “fresh”) data is inputted into the system through one sensor (such as a video camera) and immediately compared to the reference data stored in either or both data bases. As described above, a decision is made based on a predetermined algorithm. If it is determined that the fresh input data compares closely with a known hostile action or threat, an alarm is activated to summon law enforcement. Simultaneously, a recording device is activated to record the hostile event in real time.
  • accelerometers are connected to specific points of the actors' bodies. Depending on the particular actions being performed by the actors, the accelerometers may be attached to various parts of their bodies, such as the hands, lower arms, elbows, upper arms, shoulders, top of each foot, the lower leg and thigh, the neck and head. Of course other parts of the actors' bodies may similarly support an accelerometer, and some of the ones mentioned above may not be needed to record a particular action.
  • the accelerometers may be attached to the particular body joint or location using a suitable tape or adhesive and may further include a transmitter chip that transmits a signal to a multi-channel receiver located nearby, and a selected electronic filter that helps minimize transmission interference.
  • all accelerometer or a selected group may be hard wired on the actor's body and interconnected to a local master receiver.
  • the data derived from each accelerometer as the actor performs and moves his/her body includes the instantaneous acceleration of the particular body part, the change of acceleration (the jerkiness of the movement), and, through integration processing, the velocity and position at any given time.
  • JAVP These signals (collectively called “JAVP”) are processed by known mathematical operators: FFT (fast Fourier transform), cosine transform or wavelets, and then stored in a matrix format for comparison with the same processed “fresh” data, as described above.
  • FFT fast Fourier transform
  • cosine transform or wavelets are processed by known mathematical operators: FFT (fast Fourier transform), cosine transform or wavelets, and then stored in a matrix format for comparison with the same processed “fresh” data, as described above.
  • the JAVP data is collectively placed into a data base (image dictionary).
  • image dictionary includes signatures of the threat and actual assault movements of the attacker and of the response movement of the victim, paying particular attention to the movements of the attacker.
  • the weight or size of each actor is preferably taken into account. For example, ten actors representing attackers preferably vary in weight (or size) from 220 lbs. to 110 lbs. with commonly associated heights. Similarly, ten actors representing victims are selected. The twenty actors then perform a number (perhaps 100) choreographed skits or actions that factor the size difference between an attacker and a victim according to the movement of the body part, acceleration, change of acceleration, and velocity for hostile, friendly, and neutral acts. An example of an neutral act may be two people merely walking past each other without interaction.
  • JAVP data may be generated simply by recording actors performing specific actions using a conventional video sensor (such as a video camera).
  • a conventional video sensor such as a video camera
  • the same physical acts involved in the same skits or performances are carried out by the actor aggressors and actor victims, but are simply recorded by a video camera, for example.
  • the JAVP data is transformed using only image processing techniques.
  • a matrix format memory is again generated using the JAVP data and compared to each of the corresponding body part signatures derived using the accelerometers as in the above-described case.
  • similarities and the closeness of the signatures of each body part for each type of movement may be categorized: hostile (upper cut, kicking, drawing a knife, etc), friendly (shaking hands, waving, etc.), and neutral (walking past each other or standing in a line). Modifications may be made to each of these signatures in order to obtain more accurate reference signatures, according to people of different size and weight.
  • the performances by the actors would be repeated until the difference between the two signatures is understood (by the actors) and corrections made.
  • This audio-detection system includes a word-spotting/recognition and word gisting system, according to the invention, which analyzes specific words, inflections, accents, and dialects and detect spoken words and expressions that indicate hostile actions, friendly actions, or neutral ones.
  • the audio-detection system uses a shotgun-type microphone of a microphone array to achieve a high gain propagation pattern and further preferably employs appropriate noise reduction systems and common mode rejection circuitry to achieve good audio detection of the words and oral expressions provided by the attacker and the victim.
  • Word recognition and word gisting software engines are commercially available which may easily handle the relatively few words and expressions typically used during such a hostile interaction.
  • the attacker's and the victims reference words and word gisting of a hostile nature are stored in a verbal dictionary, as are those of friendly and neutral interactions.
  • An alternate approach using the above-described accelerometer technique for obtaining the reference JAVP signals associated with hostle, friendly and neutral actions is to employ doppler radar, operating at very short wavelengths, imaging radar (actually an inverse synthetic aperture radar), also operating at very short wavelengths, or laser radar. It is preferred that these active devices be operated at very low power to prevent undesireable exposure of transmitted energy to the people located within an area of transmission.
  • imaging radar actually an inverse synthetic aperture radar
  • laser radar laser radar

Abstract

A surveillance system having at least one primary video camera for translating real images of a zone into electronic video signals at a first level of resolution. The system includes means for sampling movements of an individual or individuals located within the zone from the video signal output from at least one video camera. Video signals of sampled movements of the individual is electronically compared with known characteristics of movements which are indicative of individuals having a criminal intent. The level of criminal intent of the individual or individuals is then determined and an appropriate alarm signal is produced.

Description

CROSS-REFERENCE TO COPENDING PATENT APPLICATION
This isThis application is a continuation of U.S. patent application Ser. No. 12/466,340, filed May 14, 2009, now U.S. Pat. No. Re. 42,690, which is a Reissue of U.S. patent application Ser. No. 08/898,470, filed Jul. 22, 1997, now U.S. Pat. No. 6,028,626 which is a continuation in part of patent application Ser. No. 08/367,712, filed Jan. 3, 1995, now U.S. Pat. No. 5,666,157, each of which are incorporated herein be reference in their entirety.
BACKGROUND OF THE INVENTION
A) Field of the Invention
This invention generally relates to surveillance systems, and more particularly, to trainable surveillance systems which detect and respond to specific abnormal video and audio input signals.
B) Background of the Invention
Today's surveillance systems vary in complexity, efficiency and accuracy. Earlier surveillance systems use several closed circuit cameras, each connected to a devoted monitor. This type of system works sufficiently well for low-coverage sites, i.e., areas requiring up to perhaps six cameras. In such a system, a single person could scan the six monitors, in “real” time, and effectively monitor the entire (albeit small) protected area, offering a relatively high level of readiness to respond to an abnormal act or situation observed within the protected area. In this simplest of surveillance systems, it is left to the discretion of security personnel to determine, first if there is any abnormal event in progress within the protected area, second, the level of concern placed on that particular event, and third, what actions should be taken in response to the particular event. The reliability of the entire system depends on the alertness and efficiency of the worker observing the monitors.
Many surveillance systems, however, require the use of a greater number of cameras (e.g., more than six) to police a larger area, such as at least every room located within a large museum. To adequately ensure reliable and complete surveillance within the protected area, either more personnel must be employed to constantly watch the additionally required monitors (one per camera), or fewer monitors may be used on a simple rotation schedule wherein one monitor sequentially displays the output images of several cameras, displaying the images of each camera for perhaps a few seconds. In another prior art surveillance system (referred to as the “QUAD” system), four cameras are connected to a single monitor whose screen continuously and simultaneously displays the four different images. In a “quaded quad” prior art surveillance system, sixteen cameras are linked to a single monitor whose screen now displays, continuously and simultaneously all sixteen different images. These improvements allow fewer personnel to adequately supervise the monitors to cover the larger protected area.
These improvements, however, still require the constant attention of at least one person. The above described multiple-image/single screen systems suffered from poor resolution and complex viewing. The reliability of the entire system is still dependent on the alertness and efficiency of the security personnel watching the monitors. The personnel watching the monitors are still burdened with identifying an abnormal act or condition shown on one of the monitors, determining which camera, and which corresponding zone of the protected area is recording the abnormal event, determining the level of concern placed on the particular event, and finally, determining the appropriate actions that must be taken to respond to the particular event.
Eventually, it was recognized that human personnel could not reliably monitor the “real-time” images from one or several cameras for long “watch” periods of time. It is natural for any person to become bored while performing a monotonous task, such as staring at one or several monitors continuously, waiting for something unusual or abnormal to occur; something which may never occur.
As discussed above, it is the human link which lowers the overall reliability of the entire surveillance system. U.S. Pat. No. 4,737,847 issued to Araki et al. discloses an improved abnormality surveillance system wherein motion sensors are positioned within a protected area to first determine the presence of an object of interest, such as an intruder. In the system disclosed by U.S. Pat. No. 4,737,847, zones having prescribed “warning levels” are defined within the protected area. Depending on which of these zones an object or person is detected in, moves to, and the length of time the detected object or person remains in a particular zone determines whether the object or person entering the zone should be considered an abnormal event or a threat.
The surveillance system disclosed in U.S. Pat. No. 4,737,847 does remove some of the monitoring responsibility otherwise placed on human personnel, however, such a system can only determine an intruder's “intent” by his presence relative to particular zones. The actual movements and sounds of the intruder are not measured or observed. A skilled criminal could easily determine the warning levels of obvious zones within a protected area and act accordingly; spending little time in zones having a high warning level, for example.
It is therefore an object of the present invention to provide a surveillance system which overcomes the problems of the prior art.
It is another object of the invention to provide such a surveillance system wherein a potentially abnormal event is determined by a computer prior to summoning a human supervisor.
It is another object of the invention to provide a surveillance system which compares specific measured movements of a particular person or persons with a trainable, predetermined set of “typical” movements to determine the level and type of criminal or mischievous event.
It is another object of this invention to provide a surveillance system which transmits the data from various sensors to a location where it can be recorded for evidentiary purposes. It is another object of this invention to provide such surveillance system which is operational day and night.
It is another object of this invention to provide a surveillance system which can cull out real-time events which indicate criminal intent using a weapon, by resolving the low temperature of the weapon relative to the higher body temperature and by recognizing the stances taken by the person with the weapon.
It is yet another object of this invention to provide a surveillance system which does not require “real time” observation by human personnel.
INCORPORATED BY REFERENCE
The content of the following references is hereby incorporated by reference.
    • 1. Motz L. and L. Bergstein “Zoom Lens Systems”, Journal of Optical Society of America, 3 papers in Vol. 52, 1992.
    • 2. D. G. Aviv, “Sensor Software Assessment of Advanced Earth Resources Satellite Systems”, ARC Inc. Report #70-80-A, pp2-107 through 2-119; NASA contract NAS-1-16366.
    • 3. Shio, A. and J. Sklansky “Segmentation of People in Motion”, Proc. of IEEE Workshop on Visual Motion, Princeton, N.J., October 1991.
    • 4. Agarwal, R. and J Sklansky “Estimating Optical Flow from Clustered Trajectory Velocity Time”.
    • 5. Suzuki, S. and J Sklansky “Extracting Non-Rigid Moving Objects by Temporal Edges”, IEEE, 1992, Transactions of Pattern Recognition.
    • 6. Rabiner, L. and Biing-Hwang Juang “Fundamental of Speech Recognition”, Pub. Prentice Hall, 1993, (p.434-495).
    • 7. Weibel, A. and Kai-Fu Lee Eds. “Readings in Speech Recognition”, Pub. Morgan Kaaufman, 1990 (p.267-296).
    • 8. Rabiner, L. “Speech Recognition and Speech Synthesis Systems”, Proc. IEEE, January, 1994.
SUMMARY OF THE INVENTION
A surveillance system having at least one primary video camera for translating real images of a zone into electronic video signals at a first level of resolution;
means for sampling movements of an individual or individuals located within the zone from the video signal output from at least one video camera;
means for electronically comparing the video signals of sampled movements of the individual with known characteristics of movements which are indicative of individuals having a criminal intent;
means for determining the level of criminal intent of the individual or individuals;
means for activating at least one secondary sensor and associated recording device having a second higher level of resolution, said activating means being in response to determining that the individual has a predetermined level of criminal intent.
A method for determining criminal activity by an individual within a field of view of a video camera, said method comprising:
sampling the movements of an individual located within said field of view using said video camera to generate a video signal;
electronically comparing said video signal of said with known characteristics of movements that are indicative of individuals having a criminal intent;
determining the level of criminal intent of said individual, said determining step being dependent on said electronically comparing step; and
generating a signal indicating a predetermined level of criminal intent is present as determined by said determining step.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a schematic block diagram of the video, analysis, control, alarm and recording subsystems embodying this invention;
FIG. 2A illustrates a frame K of a video camera's output of a particular environment, according to the invention, showing four representative objects (people) A, B, C, and D, wherein objects A, B and D are moving in a direction indicated with arrows, and object C is not moving;
FIG. 2B illustrates a frame K+5 of the video camera's output, according to the invention, showing objects A, B, and D are stationary, and object C is moving;
FIG. 2C illustrates a frame K+10 of the video camera's output, according to the invention, showing the current location of objects A, B, C, D, and E;
FIG. 2D illustrates a frame K+11 of the video camera's output, according to the invention, showing object B next to object C, and object E moving to the right;
FIG. 2E illustrates a frame K+12 of the video camera's output, according to the invention, showing a potential crime taking place between objects B and C;
FIG. 2F illustrates a frame K+13 of the video camera's output, according to the invention, showing objects B and C interacting;
FIG. 2G illustrates a frame K+15 of the video camera's output, according to the invention, showing object C moving to the right and object B following;
FIG. 2H illustrates a frame K+16 of the video camera's output, according to the invention, showing object C moving away from a stationary object B;
FIG. 2I illustrates a frame K+17 of the video camera's output, according to the invention, showing object B moving towards object C.
FIG. 3A illustrates a frame of a video camera's output, according to the invention, showing a “two on one” interaction of objects (people) A, B, and C;
FIG. 3B illustrates a later frame of the video camera's output of FIG. 3A, according to the invention, showing objects A and C moving towards object B;
FIG. 3C illustrates a later frame of the video camera's output of FIG. 3B, according to the invention, showing objects A and C moving in close proximity to object B;
FIG. 3D illustrates a later frame of the video camera's output of FIG. 3C, according to the invention, showing objects A and C quickly moving away from object B.
FIG. 4 is a schematic block diagram of a conventional word recognition system; and
FIG. 5 is a schematic block diagram of a video and verbal recognition system, according to the invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
Referring to FIG. 1, the basic elements of one embodiment of the invention are illustrated, including picture input means 10, which may be any conventional electronic picture pickup device operational within the infrared or visual spectrum (or both) including a vidicon and a CCD/TV camera (including the wireless type).
In another embodiment of picture input means 10, there is the deployment of a high rate camera/recorder (similar to those made by NAC Visual Systems of Woodland Hills, Calif., SONY and others). Such high rate camera/recorder systems are able to detect and record very rapid movements of body parts that are commonly indicative of a criminal intent. Such fast movements might not be resolved with a more standard 30 frames per second camera. However, most movements will be resolved with a standard 30 frames per second camera.
This picture means, may also be triggered by an alert signal from the processor of the low resolution camera or, as before, from the audio/word recognition processor when sensing a suspicious event.
In this first embodiment, the primary picture input means 10 is preferably a low cost video camera wherein high resolution is not necessary and due to the relative expense will most likely provide only moderate resolution. ((The preferred CCD/TV camera is about 1½ inches in length and about 1 inch in diameter, weighing about 3 ounces, and for particular deployment, a zoom lens attachment may be used). This device will be operating continuously and will translate the field of view (“real”) images within a first observation area into conventional video electronic signals.
In another embodiment of picture input means 10, a high rate camera/recorder, (similar to those made by NAC Visual Systems of Woodland Hills, Calif., SONY and others) is used, which would then enable the detection of even the very rapid movement of body parts that are indicative of criminal intent, and their recording. The more commonly used camera operates at 30 frames per second will be able to resolve essentially all quick body movements.
The picture input means may also be activated by an “alert” signal from the processor of the low resolution camera or from the audio/word recognition processor when sensing a suspicious event.
The picture input means for any embodiment contains a preprocessor which normalizes a wide range of illumination levels, especially for outside observation. The preprocessor to emulates a vertebrate's retina, which has an efficient and accurate normalization process. One such preprocessor (VLSI retina chip) is fabricated by the Carver Meade Laboratory of the California Institute of Technology in Pasadena, Calif. Use of this particular preprocessor chip will increase the automated vision capability of this invention whenever variation of light intensity and light reflection may otherwise weaken the picture resolution.
The signals from the picture input means 10 are converted into digitized signals and then sent to the picture processing means 12.
The processor controlling each group of cameras will be governed by an artificial intelligence system, based on dynamic pattern recognition principles, as further described below.
The picture processing means 12 includes an image raster analyzer which effectively segments each image to isolate each pair of people.
The image raster analyzer subsystem of picture processing means 12 segments each sampled image to identify and isolate each pair of objects (or people), and each “two on one” group of 3 people separately.
The “2 on 1” represents a common mugging situation in which two individuals approach a victim: one from in front of the victim and the other from behind. The forward mugger tells the potential victim that if he does not give up his money, (or watch, ring, etc.) the second mugger will shoot him, stab or otherwise harm him. The group of three people will thus be considered a potential crime in progress and will therefore be segmented and analyzed in picture processing means.
An additional embodiment of the picture means 1 is the inclusion of an optics system known as the zoom lens system. The essentials of the zoom lens subsystem are described in three papers written by L. Motz and L. Bergstein, in an article titled “Zoom Lens Systems” in the Journal of Optical Society of America, Vol. 52, April, 1992. This article is hereby incorporated by reference.
The essence of the zoom system is to vary the focal length such that an object being observed will be focused and magnified at its image plane. In an automatic version of the zoom system once an object is in the camera's field-of-view (FOV), the lens which moves to focus the object onto the camera's image plane. An error which is used to correct the focus, by the image planes's is generated by CCD array into 2 halves and measuring the difference segmenting in each until the object is at the center. Dividing the CCD array into more than 2 segments, say 4 quadrants is a way to achieve automatic centering, as is the case with mono-pulse radar. Regardless of the number of segments, the error signal is used to generate the desired tracking of the object.
In a wide field-of-view (WFOV operation, there may be more than one object, thus special attention is given to the design of the zoom system and its associated software and firmware control. Assuming 3 objects, as is the “2 on 1” potential mugging threat described above, and that the 3 persons are all in one plane, one can program a shifting from one object to the next, from one face to another face, in a prescribed sequential order. Moreover, as the objects move within the WFOV they will be automatically tracked in azimuth and elevation. In principle, the zoom would focus on the nearest object, assuming that the amount of light on each object is the same so that the prescribed sequence starting from the closes object will proceed to the remaining objects from, for example, right to left.
However, when the 3 objects are located in different planes, but still within the camera's WFOV, the zoom, with input from the segmentation subsystem of the picture analysis means 12 will focus on the object closest to the right hand side of the image plane, and then proceed to move the focus to the left, focusing on the next object and on the next sequentially.
In all of the above cases, the automatic zoom can more naturally choose to home-in on the person with the brightest emission or reflection, and then proceed to the next brightness and so forth. This would be a form of an intensity/time selection multiplex zoom system.
The relative positioning of the input camera with respect to the area under surveillance will effect the accuracy by which the image raster analyzer segments each image. In this preferred embodiment, it is beneficial for the input camera to view the area under surveillance from a point located directly above, e.g., with the input camera mounted high on a wall, a utility tower, or a traffic light support tower. The height of the input camera is preferably sufficient to minimize occlusion between the input camera and the movement of the individuals under surveillance.
Once the objects within each sampled video frame are segmented (i.e., detected and isolated), an analysis is made of the detailed movements of each object located within each particular segment of each image, and their relative movements with respect to the other objects.
Each image frame segment, once digitized, is stored in a frame by frame memory storage of section 12. Each frame from the camera input 10 is subtracted from a previous frame already stored in memory 12 using any conventional differencing process. The differencing process involving multiple differencing steps takes place in the differencing section 12. The resulting difference signal (outputted from the differencing sub-section 12) of each image indicates all the changes that have occurred from one frame to the next. These changes include any movements of the individuals located within the segment and any movements of their limbs, e.g., arms.
A collection of differencing signals for each moved object of subsequent sampled frames of images (called a “track”) allows a determination of the type, speed and direction (vector) of each motion involved and also processing which will extract acceleration, i.e., note of change of velocity: and change in acceleration with respect to time (called “jerkiness”) and will when correlating with stored signatures of known physical criminal acts. For example, subsequent differencing signals may reveal that an individual's arm is moving to a high position, such as the upper limit of that arm's motion, i.e., above his head) at a fast speed. This particular movement could be perceived, as described below, as a hostile movement with a possible criminal intent requiring the expert analysis of security personnel.
The intersection of two tracks indicates the intersection of two moved objects. The intersecting objects, in this case, could be merely the two hands of two people greeting each other, or depending on other characteristics, as described below, the intersecting objects could be interpreted as a fist of an assailant contacting the face of a victim in a less friendly greeting. In any event, the intersection of two tracks immediately requires further analysis and/or the summoning of security personnel. But the generation of an alarm, light and sound devices located, for example, on a monitor will turn a guard's attention only to that monitor, hence the labor savings. In general however, friendly interactions between individuals is a much slower physical process than is a physical assault vis-a-vis body parts of the individuals involved. Hence, friendly interactions may be easily distinguished from hostile physical acts using current low pass and high pass filters, and current pattern recognition techniques based on experimental reference data.
When a large number of sensors are distributed over a large number facilities, for example, a number of ATMs (automatic teller machines), associated with particular bank branches and in a particular state or states and all operated under a single bank network control on a time division multiplexed basis, then only a single monitor is required.
A commercially available software tool may enhance object-movement analysis between frames (called optical flow computation). (see ref. 3 and 4) With optical flow computation, specific (usually bright) reflective elements, called farkles, emitted from the clothing and/or the body parts of an individual of one frame are subtracted from a previous frame. The bright portions will inherently provide sharper detail and therefore will yield more accurate data regarding the velocities of the relative moving objects. Additional computation, as described below, will provide data regarding the acceleration and even change in acceleration or “jerkiness” of each moving part sampled.
The physical motions of the individuals involved in an interaction, will be detected by first determining the edges of the of each person imaged. And the movements of the body parts will then be observed by noting the movements of the edges of the body parts of the (2 or 3) individuals involved in the interaction.
The differencing process will enable the determination of the velocity and acceleration and rate of acceleration of those body parts.
The now processed signal is sent to comparison means 14 which compares selected frames of the video signals from the picture input means 10 with “signature” video signals stored in memory 16. The signature signals are representative of various positions and movements of the body ports of an individual having various levels of criminal intent. The method for obtaining the data base of these signature video signals in accordance with another aspect of the invention is described in greater detail below.
If a comparison is made positive with one or more of the signature video signals, an output “alert” signal is sent from the comparison means 14 to a controller 18. The controller 18 controls the operation of a secondary, high resolution picture input means (video camera) 20 and a conventional monitor 22 and video recorder 24. The field of view of the secondary camera 20 is preferably at most, the same as the field of view of the primary camera 10, surveying a second observation area. The recorder 24 may be located at the site and/or at both a law enforcement facility (not shown) and simultaneously at a Court office or legal facility to prevent loss of incriminating information due to tampering.
The purpose of the secondary camera 20 is to provide a detailed video signal of the individual having assumed criminal intent and also to improve false positive and false negative performance. This information is recorded by the video recorder 24 and displayed on a monitor 22. An alarm bell or light (not shown) or both may be provided and activated by an output signal from the controller 20 to summon a supervisor to immediately view the pertinent video images showing the apparent crime in progress and access its accuracy.
In still another embodiment of the invention, a VCR 26 is operating continuously (using a 6 hour loop-tape, for example). The VCR 26 is being controlled by the VCR controller 28. All the “real-time” images directly from the picture input means 10 are immediately recorded and stored for at least 6 hours, for example. Should it be determined that a crime is in progress, a signal from the controller 18 is sent to the VCR controller 28 changing the mode of recording from tape looping mode to non-looping mode. Once the VCR 26 is changed to a non-looping mode, the tape will not re-loop and will therefore retain the perhaps vital recorded video information of the surveyed site, including the crime itself, and the events leading up to the crime.
When the non-looping mode is initiated, the video signal may also be transmitted to a VCR located elsewhere; for example, at a law enforcement facility and, simultaneously to other secure locations of the Court and its associated offices.
Prior to the video signals being compared with the “signature” signals stored in memory, each sampled frame of video is “segmented” into parts relating to the objects detected therein. To segment a video signal, the video signal derived from the vidicon or CCD/TV camera is analyzed by an image raster analyzer. Although this process causes slight signal delays, it is accomplished nearly in real time.
At certain sites, or in certain situations, a high resolution camera may not be required or otherwise used. For example, the resolution provided by a relatively simple and low cost camera may be sufficient. Depending on the level of security for the particular location being surveyed, and the time of day, the length of frame intervals between analyzed frames may vary. For example, in a high risk area, every frame from the CCD/TV camera may be analyzed continuously to ensure that the maximum amount of information is recorded prior to and during a crime. In a low risk area, it may be preferred to sample perhaps every 10 frames from each camera, sequentially. If, during such a sampling, it is determined that an abnormal or suspicious event is occurring, such as two people moving very close to each other, then the system would activate an alert mode wherein the system becomes “concerned and curious” in the suspicious actions and the sampling rate is increased to perhaps every 5 frames or even every frame. As described in greater detail below, depending on the type of system employed (i.e., video only, audio only or both), during such an alert mode, the entire system may be activated wherein both audio and video system begin to sample the environment for sufficient information to determine the intent of the actions.
Referring to FIG. 2, several frames of a particular camera output are shown to illustrate the segmentation process performed in accordance with the invention. The system begins to sample at frame K and determines that there are four objects (previously determined to be people, as described below), A-D located within a particular zone being policed. Since nothing unusual is determined from the initial analysis, the system does not warrant an “alert” status. People A, B, and D are moving according to normal, non-criminal intent, as could be observed.
A crime likelihood is indicated when frames K+10 through K+13 are analyzed by the differencing process. And if the movement of the body parts indicate velocity, acceleration and “jerkiness” that compare positively with the stored digital signals depicting movements of known criminal physical assaults, it is likely that a crime is in progress here.
Additionally, if a large velocity of departure is indicated when person C moves away from person B, as indicated in frames K+15 through K+17, a larger level of confidence, is attained in deciding that a physical criminal act has taken plate or is about to.
An alarm is generated the instant any of the above conditions is established. This alarm condition will result in sending in Police or Guards to the crime site, activating the high resolution CCD/TV camera to record the face of the person committing the assault, a loud speaker being activated automatically, playing a recorded announcement warning the perpetrator the seriousness of his actions now being undertaken and demanding that he cease the criminal act. After dark a strong light will be turned on automatically. The automated responses will be actuated the instant an alarm condition is adjudicated by the processor. Furthermore, an alarm signal is sent to the police station and the same video signal of the event, is transmitted to a court appointed data collection office, to the Public Defender's office and the District Attorney's Office.
As described above, it is necessary to compare the resulting signature of physical body parts motion involved in a physical criminal act, that is expressed by specific motion characteristics (i.e., velocity, acceleration, change of acceleration), with a set of signature files of physical criminal acts, in which body parts motion are equally involved. This comparison, is commonly referred to as pattern matching and is part of the pattern recognition process.
The files of physical criminal acts, which involve body parts movements such as hands, arms, elbows, shoulder, head, torso, legs, and feet we obtained, a priority, by experiments and simulations of physical criminal acts gathered from “dramas” that are enacted by professional actors, the data gathered from experienced muggers who have been caught by the police as well as victims who have reported details of their experiences will help the actors perform accurately. Video of their motions involved in these simulated acts will be stored in digitized form and files prepared for each of the body parts involved, in the simulated physical criminal acts.
The present invention could be easily implemented at various sites to create effective “Crime Free” zones. In another embodiment, the above described Abnormality Detection System includes an RF-ID (Radio Frequency Identification) tag, to assist in the detection and tracking of individuals within the field of view of a camera.
I.D. cards or tags are worn by authorized individuals. The tags response when queried by the RF Interrogator. The response signal of the tags propagation pattern which is adequately registered with the video sensor. The “Tags” are sensed in video are assumed friendly and authorized. This information will simplify the segmentation process.
A light connected to each RF-ID card will be turned ON, when a positive response to an interrogation signal is established. The light will appear on the computer generated grid (also on the screen of the monitor) and the intersection of tracks clearly indicated, followed by their physical interaction. But also noted will be the intersection between the tagged and the untagged individuals. In all of such cases, the segmentation process will be simpler.
There are many manufacturers of RF-ID cards and Interrogators, three major ones are, The David Sarnoff Research Center of Princeton, N.J., AMTECH of Dallas, Tex. and MICRON Technology of Boise, Id.
The applications of the present invention include stationary facilities: banks and ATMs, hotels, private residence halls and dormitories, high rise and low rise office and residential buildings, public and private schools from kindergarten through high-school, colleges and universities, hospitals, sidewalks, street crossing, parks, containers and container loading areas, shipping piers, train stations, truck loading stations, airport passenger and freight facilities, bus stations, subway stations, move houses, theaters, concert halls and arenas, sport arenas, libraries, churches, museums, stores, shopping malls, restaurants, convenience stores, bars, coffee shops, gasoline stations, highway rest stops, tunnels, bridges, gateways, sections of highways, toll booths, warehouses, and depots, factories and assembly rooms, law enforcement facilities including jails.
Further applications of the invention include areas of moving platforms: automobiles, trucks, buses, subway cars, train cars, freight and passenger, boats and ships (passenger and freight, tankers, service vehicles, construction vehicles, on and off-road, containers and their carriers, and airplanes. And also in military applications that will include but will not be limited to assorted military ground, sea, and air mobile vehicles and assorted military ground, sea, and air mobile vehicles and platforms as well as stationary facilities where the protection of low, medium, and high value targets are necessary; such targets are common in the military but have equivalents in the civilian areas wherein this invention will serve both sectors.
As a deterrence to car-jacking a tiny CCD/TV camera connected surreptitiously at the ceiling of the car, or in the rear-view mirror, through a pin hole lens and focused at the driver's seat, will be connected to the video processor to record the face of the drive. The camera is triggered by the automatic word recognition processor that will identify the well known expressions commonly used by the car-jacker. The video picture will be recorded and then transmitted via cellular phone in the car. Without a phone, the short video recording of the face of the car-jacker will be held until the car is found by the police, but now with the evidence (the picture of the car-jacker) in hand.
In this present surveillance system, the security personnel manning the monitors are alerted only to video images which show suspicious actions (criminal activities) within a prescribed observation zone. The security personnel are therefore used to access the accuracy of the crime and determine the necessary actions for an appropriate response. By using computers to effectively filter out all normal and noncriminal video signals from observation areas, fewer security personnel are required to survey and “secure” a greater overall area (including a greater number of observation areas, i.e., cameras).
It is also contemplated that the present system could be applied to assist blind people “see”. A battery operated portable version of the video system would automatically identify known objects in its field of view and a speech synthesizer would “say” the object. For example, “chair”, “table”, etc. would indicate the presence of a chair and a table.
Depending on the area to be policed, it is preferable that at least two and perhaps three cameras (or video sensors) are used simultaneously to cover the area. Should one camera sense a first level of criminal action, the other two could be manipulated to provide a three dimensional perspective coverage of the action. The three dimensional image of a physical interaction in the policed area would allow observation of a greater number of details associated with the steps: accost, threat, assault, response and post response. The conversion from the two dimensional image to the three dimensional image is known as “random transform”.
In the extended operation phase of the invention as more details of the physical variation of movement characteristics of physical threats and assaults against a victim and also the speaker independent (male, female of different ages groups) and dialect independent words and terse sentences, with corresponding responses, will enable automatic recognition of a criminal assault, without he need of guard, unless required by statutes and other external requirements.
In another embodiment of the present invention, both video and acoustic information is sampled and analyzed. The acoustic information is sampled and analyzed in a similar manner to the sampling and analyzing of the above-described video information. The audio information is sampled and analyzed in a manner shown in FIG. 4, and is based on prior art. (references 6 and 7).
The employment of the audio speech band, with its associated Automatic Speech Recognition (ASR) system, will not only reduce the false alarm rate resulting from the video analysis, but can also be used to trigger the video and other sensors if the sound threat predates the observed threat.
Referring to FIG. 4, a conventional automatic word recognition system is shown, including an input microphone system 40, an analysis subsystem 42, a template subsystem 44, a pattern comparator 46, and a post-processor and decision logic subsystem 48.
In operation, upon activation, the acoustic/audio policing system will begin sampling all (or a selected portion) of nearby acoustic signals. The acoustic signals will include voices and background noise. The background noise signals are generally known and predictable, and may therefore be easily filtered out using conventional filtering techniques. Among the expected noise signals are unfamiliar speech, automotive related sounds, honking, sirens, the sound of wind and/or rain.
The microphone input system 40 pick-up the acoustic signals and immediately filter out the predictable background noise signals and amplify the remaining recognizable acoustic signals. The filtered acoustic signals are analyzed in the analysis subsystem 42 which processes the signals by means of digital and spectral analysis techniques. The output of the analysis subsystem is compared in the pattern comparater subsystem 46 with selected predetermined words stored in memory in 44. The post processing and decision logic subsystem 48 generates an alarm signal, as described below.
The templates 44 include perhaps about 100 brief and easily recognizable terse expressions, some of which are single words, and are commonly used by those intent on a criminal act. Some examples of commonly used word phrases spoken by a criminal to a victim prior to a mugging, for example, include: “Give me your money”, “This is a stick-up”, “Give me your wallet and you won't get hurt” . . . etc. Furthermore, commonly used replies from a typical victim during such a mugging may also be stored as template words, such as “help”, and certain sounds such as shrieks, screams and groans, etc.
The specific word templates, from which inputted acoustic sounds are compared with, must be chosen carefully, taking into account the particular accents and slang of the language spoken in the region of concern (e.g., the southern cities of the U.S. will require a different template 44 than the one used for a recognition system in the New York City region of the U.S.).
The output of the word recognition system shown in FIG. 4 is used as a trigger signal to activate a sound recorder, or a camera used elsewhere in the invention, as described below.
The preferred microphone used in the microphone input subsystem 40 is a shotgun microphone, such as those commercially available from the Sennheiser Company of Frankfurt, Germany. These microphone have a super-cardioid propagation pattern. However, the gain of the pattern may be too small for high traffic areas and may therefore require more than one microphone in an array configuration to adequately focus and track in these areas. The propagation pattern of the microphone system enables better focusing on a moving sound source (e.g., a person walking and talking). A conventional directional microphone may also be used in place of a shot-gun type microphone, such as those made by the Sony Corporation of Tokyo, Japan. Such directional microphones will achieve similar gain to the shot-gun type microphones, but with a smaller physical structure.
A feedback loop circuit (not specifically shown) originating in the post processing subsystem 48 will direct the microphone system to track a particular dynamic source of sound within the area surveyed by video cameras.
An override signal from the video portion of the present invention will activate and direct the microphone system towards the direction of the field of view of the camera. In other words, should the video system detect a potential crime in progress, the video system will control the audio recording system towards the scene of interest. Likewise, should the audio system detect words of an aggressive nature, as described above, the audio system will direct appropriate video cameras to visually cover and record the apparent source of the sound.
A number of companies have developed very accurate and efficient, speaker independent word recognition systems based on a hidden Markov model (HMM) in combination with an artificial neural network (ANN). These companies include IBM of Armonk, N.Y., AT&T Bell Laboratories, Kurtzweil of Cambridge, Mass. and Lernout and Hauspie of Belgium.
Put briefly, the HMM system uses probability statistics to predict a particular spoken word following recognition of a primary word unit, syllable or phoneme. For example, as the word “money” is inputted into an HMM word recognition system, the first recognized portion of the word is “mon . . . ”. The HMM system immediately recognizes this word stem and determines that the spoken word could be “MONDAY”, “MONopoly”, or “MONey”, etc. The resulting list of potential words is considerably shorter than the entire list of all spoken words of the English language. Therefore, the HMM system employed with the present invention allows both the audio and video systems to operate quickly and use HMM probability statistics to predict future movements or words based on an early recognition of initial movements and word stems.
The HMM system may be equally employed in the video recognition system. For example, if a person's arm quickly moves above his head, the HMM system may determine that there is a high probability that the arm will quickly come down, perhaps indicating a criminal intent.
The above-described system actively compares input data signals from a video camera, for example, with known reference data of specific body movements stored in memory. In accordance with the invention, a method of obtaining the “reference data” (or ground truth data) is described. This reference data describes threats, actual criminal physical acts, verbal threats and verbal assaults, and also friendly physical acts and friendly words, and neutral interactions between interacting people.
According to the invention, the reference data may be obtained using any of at least the following described three methods including a) attaching accelerometers at predetermined points (for example arm and leg joints, hips, and the forehead) of actors; b) using a computer to derive 3-D models of people (stored in the computer's memory as pixel data) and analyze the body part movements of the people; and c) scanning (or otherwise downloading) video data from movie and TV clips of various physical and verbal interactions into a computer to analyze specific movements and sounds.
While the above-identified three approaches should yield similar results, the preferred method for obtaining reference data is includes attaching accelerometers to actors while performing various actions or “events” of interest: abnormal (e.g., criminal or generally quick, violent movements), normal (e.g., shaking hands, slow and smooth movements), and neutral behavior (e.g., walking).
In certain environments, in particular where many people are moving in different directions, such as during rush hour in the concourse of Grand Central Station or in Central Park, both located in New York City, it may prove very difficult to analyze the specific movements of each person located within the field of view of a surveillance camera. To overcome the analyzing burden in these environments, according to another embodiment of the invention, the people located within the environment are provided personal ID cards that include an electronic radio frequency (rf) transmitter. The transmitter of each radio-frequency identification card (RFID) transmits an rf signal that identifies the person carrying the card. Receivers located in the area of a surveillance camera can receive the identification information and use it to help identify the different people located within the field of the near by surveillance camera (or microphone, in the case of audio analysis). In one possible arrangement, people may be issued an RFID card prior to entering a particular area, such as a U.S. Tennis Open event. In such instance, a clearance check would be made for each person prior to them receiving such a card. Once within the secure area, surveillance cameras would associate card-holders as less likely to cause trouble and would be suspicious of anyone within the field of the camera's view not being identified by an RFID card.
As described above, the basic configuration of the invention (as shown in FIGS. 1 and 2) uses video and audio sensors (such as, respectively, a camera and a microphone), and potentially other active and passive sensing and processing devices and systems (including the use of radar and ladar and other devices that operate in all areas of the electromagnetic spectrum) to detect threats and actual criminal acts occurring with a field of view of a camera (a video sensor). The system described above, and according to the invention, initially requires the collection of “reference values” which correspond to specific known acts of threat, actual assault (both physical and verbal), and other physical and verbal interactions that are considered friendly or neutral. Video components of recorded “reference data” is stored in a physical movement dictionary (or data base), while audio components of such reference data is stored in a verbal utterance dictionary (or data base).
In operation of the earlier described system, real time (or “fresh”) data is inputted into the system through one sensor (such as a video camera) and immediately compared to the reference data stored in either or both data bases. As described above, a decision is made based on a predetermined algorithm. If it is determined that the fresh input data compares closely with a known hostile action or threat, an alarm is activated to summon law enforcement. Simultaneously, a recording device is activated to record the hostile event in real time.
The above-described reference data is preferably obtained through the use of actors performing specific movements of hostility, threats, and friendly and neutral actions and other actors performing neutral actions of greetings and also simulating a victim's response to acts of aggression, hostility and friendship. According to the invention, accelerometers are connected to specific points of the actors' bodies. Depending on the particular actions being performed by the actors, the accelerometers may be attached to various parts of their bodies, such as the hands, lower arms, elbows, upper arms, shoulders, top of each foot, the lower leg and thigh, the neck and head. Of course other parts of the actors' bodies may similarly support an accelerometer, and some of the ones mentioned above may not be needed to record a particular action.
The accelerometers may be attached to the particular body joint or location using a suitable tape or adhesive and may further include a transmitter chip that transmits a signal to a multi-channel receiver located nearby, and a selected electronic filter that helps minimize transmission interference. Alternatively, all accelerometer or a selected group may be hard wired on the actor's body and interconnected to a local master receiver. The data derived from each accelerometer as the actor performs and moves his/her body, includes the instantaneous acceleration of the particular body part, the change of acceleration (the jerkiness of the movement), and, through integration processing, the velocity and position at any given time. These signals (collectively called “JAVP”) are processed by known mathematical operators: FFT (fast Fourier transform), cosine transform or wavelets, and then stored in a matrix format for comparison with the same processed “fresh” data, as described above. The JAVP data is collectively placed into a data base (image dictionary). The image dictionary includes signatures of the threat and actual assault movements of the attacker and of the response movement of the victim, paying particular attention to the movements of the attacker.
In making the “reference data”, the weight or size of each actor is preferably taken into account. For example, ten actors representing attackers preferably vary in weight (or size) from 220 lbs. to 110 lbs. with commonly associated heights. Similarly, ten actors representing victims are selected. The twenty actors then perform a number (perhaps 100) choreographed skits or actions that factor the size difference between an attacker and a victim according to the movement of the body part, acceleration, change of acceleration, and velocity for hostile, friendly, and neutral acts. An example of an neutral act may be two people merely walking past each other without interaction.
Once an initial set of JAVP data is generated through the use of actors carrying accelerometers, as described above, further JAVP data may be generated simply by recording actors performing specific actions using a conventional video sensor (such as a video camera). In this case, the same physical acts involved in the same skits or performances are carried out by the actor aggressors and actor victims, but are simply recorded by a video camera, for example. The JAVP data is transformed using only image processing techniques. A matrix format memory is again generated using the JAVP data and compared to each of the corresponding body part signatures derived using the accelerometers as in the above-described case. In doing this, similarities and the closeness of the signatures of each body part for each type of movement may be categorized: hostile (upper cut, kicking, drawing a knife, etc), friendly (shaking hands, waving, etc.), and neutral (walking past each other or standing in a line). Modifications may be made to each of these signatures in order to obtain more accurate reference signatures, according to people of different size and weight.
If the differences between the video-only JAVP data and the accelerometer JAVP data is more than a predetermined amount, the performances by the actors would be repeated until the difference between the two signatures is understood (by the actors) and corrections made.
The difference between the accelerometer and video sensor signatures based on input of same physical movements, bounds the range of incremental change for the reference signatures.
Typically accompanying each of the hostile, friendly, and neutral acts performed by the actors, spoken words and expressions are verbalized by the attacker and by the victim. This audio-detection system includes a word-spotting/recognition and word gisting system, according to the invention, which analyzes specific words, inflections, accents, and dialects and detect spoken words and expressions that indicate hostile actions, friendly actions, or neutral ones.
The audio-detection system uses a shotgun-type microphone of a microphone array to achieve a high gain propagation pattern and further preferably employs appropriate noise reduction systems and common mode rejection circuitry to achieve good audio detection of the words and oral expressions provided by the attacker and the victim.
Word recognition and word gisting software engines are commercially available which may easily handle the relatively few words and expressions typically used during such a hostile interaction. The attacker's and the victims reference words and word gisting of a hostile nature are stored in a verbal dictionary, as are those of friendly and neutral interactions.
Referring to FIG. 5, in operation, according to this embodiment of the invention, physical movements and verbal utterances of people in a field of view of an area under surveillance are recorded by an appropriate video camera and microphone. Image data from the camera is processed (e.g., filtered), as described above and compared to image data stored within the reference image dictionary, which is compiled in a manner described above. Similarly, audio information from the microphone is processed (filtered) and compared with known verbal utterances from the reference verbal dictionary, which is compiled in a manner described above.
If either an image or a verbal utterance matches (to a predetermined degree) a known image or verbal utterance of hostility, then an alarm is activated and recording equipment is turned on.
An alternate approach using the above-described accelerometer technique for obtaining the reference JAVP signals associated with hostle, friendly and neutral actions is to employ doppler radar, operating at very short wavelengths, imaging radar (actually an inverse synthetic aperture radar), also operating at very short wavelengths, or laser radar. It is preferred that these active devices be operated at very low power to prevent undesireable exposure of transmitted energy to the people located within an area of transmission. Among the benefits of using any of the above-listed active sensors is their ability to detect and analyze movements of selected body parts at a distance, in darkness (e.g., at night), and depending on the range, through inclement weather.

Claims (24)

What is claimed is:
1. A method for determining criminal activity by an individual within a field of view of a video camera, said method comprising:
sampling the movements of an individual located within said field of view using said video camera to generate a video signal;
electronically comparing said video signal of said video camera with known characteristics of movements that are indicative of an individual having criminal intent;
determining the level of criminal intent of said individual, said determining step being dependent on said electronically comparing step; and
generating a signal indicating that a predetermined level of criminal intent is present as determined by said determining step.
2. A method for determining criminal activity by an individual within a field of view of at least one video camera, the method comprising:
generating, using said at least one video camera, at least one video signal of the individual within the field of view of the at least one video camera;
sampling a relative movement, from one or more images captured by said at least one video camera of said field of view, of the individual with respect to a moved, movable or moving object captured by said at least one video camera of said field of view;
electronically comparing the sampled relative movement of the individual with known characteristics of movements that are indicative of an individual having criminal intent;
determining a level of criminal intent of the individual based on the compared sampled movement of the individual; and
generating a signal indicating that a predetermined level of criminal intent is present if the determined level of criminal intent of the individual establishes that the predetermined level of criminal intent is present.
3. The method according to claim 2, wherein the relative movement of the individual with respect to the object comprises an arm movement, a leg movement, an arm joint movement, a leg joint movement, an elbow movement, a shoulder movement, a head movement a torso movement, a hand movement, a foot movement, or combinations thereof, of the individual.
4. The method according to claim 3, wherein sampling the relative movement of the individual with respect to the object further comprises sampling an edge of an arm, a leg, an elbow, a shoulder, a head, a torso, a hand, a foot, or combinations thereof, of the individual.
5. The method according to claim 2, wherein electronically comparing the sampled relative movement of the individual with respect to the object with known characteristics of movements that are indicative of an individual having criminal intent comprises correlating a track of the sampled relative movement to known characteristics of movements that are indicative of an individual having criminal intent.
6. The method according to claim 2, wherein electronically comparing the sampled relative movement of the individual with respect to an object with known characteristics of movements that are indicative of an individual having criminal intent comprises pattern matching of the sampled relative movement to known movements that are indicative of an individual having criminal intent.
7. The method according to claim 2, wherein determining the level of criminal intent of the individual comprises detecting a speed, a direction, a jerkiness, or combinations thereof, of the sampled relative movements of the individual with respect to the object.
8. The method according to claim 7, wherein determining the level of criminal intent of the individual further comprises detecting a change of velocity, a change in acceleration, a jerkiness, or combinations thereof, of the sampled relative movements of the individual with respect to the object.
9. The method according to claim 2, wherein determining the level of criminal intent of the individual further comprises detecting a change of velocity, a change in acceleration, a jerkiness, or combinations thereof, of the sampled relative movements of the individual with respect to the object.
10. The method according to claim 2, wherein the video signal of the individual generated within the field of view of the video camera comprises a first resolution,
the method further comprising generating a second video signal of the individual within the field of view of a second video camera, the video signal comprising a second resolution, the first resolution being lower than the second resolution.
11. The method according to claim 2, further comprising:
generating an audio signal of the individual;
sampling the audio signal of the individual; and
electronically comparing the sampled audio signal of the individual with known characteristics of sounds that are indicative of an individual having criminal intent, and
wherein determining the level of criminal intent of the individual is further based on a result of electronic comparing of the sampled audio signal of the individual with the known characteristics of sounds that are indicative of an individual having criminal intent.
12. The method according to claim 11, wherein determining the level of criminal intent of the individual comprises detecting a speed, a direction, a jerkiness, a change of velocity, a change in acceleration, or a combination thereof, of the sampled relative movements of the individual with respect to the object.
13. The method according to claim 11, wherein the video signal of the individual generated within the field of view of the video camera comprises a first resolution,
the method further comprising generating a second video signal of the individual within the field of view of a second video camera, the video signal comprising a second resolution, the first resolution being lower than the second resolution.
14. The method according to claim 2, wherein determining the level of criminal intent of the individual further comprises detecting one or more of a recognized word and a recognized expression.
15. The method according to claim 3, wherein the object comprises at least one body part of the individual or at least one identified object.
16. The method according to claim 15, wherein the at least one body part of the individual comprises a hand, an arm, an elbow, a shoulder, a head, a torso, a leg or a foot.
17. The method according to claim 15, wherein the object comprises a weapon.
18. The method according to claim 2, further comprising controlling a second video camera in response to the signal indicating that a predetermined level of criminal intent is present.
19. The method according to claim 2, wherein the sampled relative movement of the individual with respect to the object comprises a movement of the object with respect to the individual, a lack of a movement of the object with respect to the individual, a jerkiness of motion of the object with respect to the individual, or a jerkiness of motion of an individual with respect to the object, or combinations thereof.
20. The method according to claim 2, further comprising sensing the relative movement of the individual using ladar or radar.
21. The method according to claim 2, wherein determining the level of criminal intent of the individual further comprises determining the temperature difference between the individual and the object.
22. The method according to claim 2, wherein sampling the relative movement of the individual with respect to the object further comprises sampling an edge of an arm, a leg, an elbow, a shoulder, a head, a torso, a hand, a foot, or combinations thereof, of the individual.
23. The method of claim 2, further comprising a segmentation step.
24. The method of claim 2, wherein the known characteristics of movements that are indicative of the individual having criminal intent are reference values.
US13/230,490 1995-01-03 2011-09-12 Abnormality detection and surveillance system Expired - Lifetime USRE44225E1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/230,490 USRE44225E1 (en) 1995-01-03 2011-09-12 Abnormality detection and surveillance system

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US08/367,712 US5666157A (en) 1995-01-03 1995-01-03 Abnormality detection and surveillance system
US08/898,470 US6028626A (en) 1995-01-03 1997-07-22 Abnormality detection and surveillance system
US12/466,340 USRE42690E1 (en) 1995-01-03 2009-05-14 Abnormality detection and surveillance system
US13/230,490 USRE44225E1 (en) 1995-01-03 2011-09-12 Abnormality detection and surveillance system

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US08/898,470 Reissue US6028626A (en) 1995-01-03 1997-07-22 Abnormality detection and surveillance system

Publications (1)

Publication Number Publication Date
USRE44225E1 true USRE44225E1 (en) 2013-05-21

Family

ID=25409505

Family Applications (3)

Application Number Title Priority Date Filing Date
US08/898,470 Ceased US6028626A (en) 1995-01-03 1997-07-22 Abnormality detection and surveillance system
US12/466,340 Expired - Lifetime USRE42690E1 (en) 1995-01-03 2009-05-14 Abnormality detection and surveillance system
US13/230,490 Expired - Lifetime USRE44225E1 (en) 1995-01-03 2011-09-12 Abnormality detection and surveillance system

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US08/898,470 Ceased US6028626A (en) 1995-01-03 1997-07-22 Abnormality detection and surveillance system
US12/466,340 Expired - Lifetime USRE42690E1 (en) 1995-01-03 2009-05-14 Abnormality detection and surveillance system

Country Status (3)

Country Link
US (3) US6028626A (en)
EP (1) EP0997041A1 (en)
WO (1) WO1999005867A1 (en)

Cited By (63)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120268269A1 (en) * 2011-04-19 2012-10-25 Qualcomm Incorporated Threat score generation
US9084411B1 (en) 2014-04-10 2015-07-21 Animal Biotech Llc Livestock identification system and method
US20160133122A1 (en) * 2013-06-25 2016-05-12 Nec Corporation Sensitivity adjustment device, sensitivity adjustment method, storage medium, and monitoring system
US20170262438A1 (en) * 2005-10-26 2017-09-14 Cortica, Ltd. System and method for determining analytics based on multimedia content elements
US20170275134A1 (en) * 2014-11-26 2017-09-28 Otis Elevator Company Elevator security and control system based on passenger movement
US9984154B2 (en) 2015-05-01 2018-05-29 Morpho Detection, Llc Systems and methods for analyzing time series data based on event transitions
US20180300887A1 (en) * 2010-01-11 2018-10-18 Shoppertrak Rct Corporation System and process for detecting, tracking and counting human objects of interest
US10559189B1 (en) 2018-12-14 2020-02-11 John Dante Menna System, method and apparatus for providing voice alert notification of an incident
US10699421B1 (en) 2017-03-29 2020-06-30 Amazon Technologies, Inc. Tracking objects in three-dimensional space using calibrated visual cameras and depth cameras
US10742340B2 (en) 2005-10-26 2020-08-11 Cortica Ltd. System and method for identifying the context of multimedia content elements displayed in a web-page and providing contextual filters respective thereto
US10748038B1 (en) 2019-03-31 2020-08-18 Cortica Ltd. Efficient calculation of a robust signature of a media unit
US10789535B2 (en) 2018-11-26 2020-09-29 Cartica Ai Ltd Detection of road elements
US10839694B2 (en) 2018-10-18 2020-11-17 Cartica Ai Ltd Blind spot alert
US10839203B1 (en) 2016-12-27 2020-11-17 Amazon Technologies, Inc. Recognizing and tracking poses using digital imagery captured from multiple fields of view
US10848590B2 (en) 2005-10-26 2020-11-24 Cortica Ltd System and method for determining a contextual insight and providing recommendations based thereon
US10846544B2 (en) 2018-07-16 2020-11-24 Cartica Ai Ltd. Transportation prediction system and method
US10902049B2 (en) 2005-10-26 2021-01-26 Cortica Ltd System and method for assigning multimedia content elements to users
US10949773B2 (en) 2005-10-26 2021-03-16 Cortica, Ltd. System and methods thereof for recommending tags for multimedia content elements based on context
US11019161B2 (en) 2005-10-26 2021-05-25 Cortica, Ltd. System and method for profiling users interest based on multimedia content analysis
US11030442B1 (en) * 2017-12-13 2021-06-08 Amazon Technologies, Inc. Associating events with actors based on digital imagery
US11029685B2 (en) 2018-10-18 2021-06-08 Cartica Ai Ltd. Autonomous risk assessment for fallen cargo
US11032017B2 (en) 2005-10-26 2021-06-08 Cortica, Ltd. System and method for identifying the context of multimedia content elements
US11037015B2 (en) 2015-12-15 2021-06-15 Cortica Ltd. Identification of key points in multimedia data elements
US11061933B2 (en) 2005-10-26 2021-07-13 Cortica Ltd. System and method for contextually enriching a concept database
US11126869B2 (en) 2018-10-26 2021-09-21 Cartica Ai Ltd. Tracking after objects
US11132548B2 (en) 2019-03-20 2021-09-28 Cortica Ltd. Determining object information that does not explicitly appear in a media unit signature
US11170647B2 (en) 2019-02-07 2021-11-09 Cartica Ai Ltd. Detection of vacant parking spaces
US11195043B2 (en) 2015-12-15 2021-12-07 Cortica, Ltd. System and method for determining common patterns in multimedia content elements based on key points
US11216498B2 (en) 2005-10-26 2022-01-04 Cortica, Ltd. System and method for generating signatures to three-dimensional multimedia data elements
US11233956B2 (en) 2020-03-31 2022-01-25 Western Digital Technologies, Inc. Sensor system with low power sensor devices and high power sensor devices
US11232294B1 (en) 2017-09-27 2022-01-25 Amazon Technologies, Inc. Generating tracklets from digital imagery
US11232326B2 (en) 2007-07-03 2022-01-25 Shoppertrak Rct Corporation System and process for detecting, tracking and counting human objects of interest
US11284041B1 (en) 2017-12-13 2022-03-22 Amazon Technologies, Inc. Associating items with actors based on digital imagery
US11285963B2 (en) 2019-03-10 2022-03-29 Cartica Ai Ltd. Driver-based prediction of dangerous events
US11361014B2 (en) 2005-10-26 2022-06-14 Cortica Ltd. System and method for completing a user profile
US11386139B2 (en) 2005-10-26 2022-07-12 Cortica Ltd. System and method for generating analytics for entities depicted in multimedia content
US11392738B2 (en) 2018-10-26 2022-07-19 Autobrains Technologies Ltd Generating a simulation scenario
US11398094B1 (en) 2020-04-06 2022-07-26 Amazon Technologies, Inc. Locally and globally locating actors by digital cameras and machine learning
US11403336B2 (en) 2005-10-26 2022-08-02 Cortica Ltd. System and method for removing contextually identical multimedia content elements
US11443516B1 (en) 2020-04-06 2022-09-13 Amazon Technologies, Inc. Locally and globally locating actors by digital cameras and machine learning
US11468698B1 (en) 2018-06-28 2022-10-11 Amazon Technologies, Inc. Associating events with actors using digital imagery and machine learning
US11468681B1 (en) 2018-06-28 2022-10-11 Amazon Technologies, Inc. Associating events with actors using digital imagery and machine learning
US11482045B1 (en) 2018-06-28 2022-10-25 Amazon Technologies, Inc. Associating events with actors using digital imagery and machine learning
US11537636B2 (en) 2007-08-21 2022-12-27 Cortica, Ltd. System and method for using multimedia content as search queries
US11590988B2 (en) 2020-03-19 2023-02-28 Autobrains Technologies Ltd Predictive turning assistant
US11593662B2 (en) 2019-12-12 2023-02-28 Autobrains Technologies Ltd Unsupervised cluster generation
US11604847B2 (en) 2005-10-26 2023-03-14 Cortica Ltd. System and method for overlaying content on a multimedia content element based on user interest
US11613261B2 (en) 2018-09-05 2023-03-28 Autobrains Technologies Ltd Generating a database and alerting about improperly driven vehicles
US11620327B2 (en) 2005-10-26 2023-04-04 Cortica Ltd System and method for determining a contextual insight and generating an interface with recommendations based thereon
US11643005B2 (en) 2019-02-27 2023-05-09 Autobrains Technologies Ltd Adjusting adjustable headlights of a vehicle
US11694088B2 (en) 2019-03-13 2023-07-04 Cortica Ltd. Method for object detection using knowledge distillation
US11704292B2 (en) 2019-09-26 2023-07-18 Cortica Ltd. System and method for enriching a concept database
US11727056B2 (en) 2019-03-31 2023-08-15 Cortica, Ltd. Object detection based on shallow neural network that processes input images
US11741687B2 (en) 2019-03-31 2023-08-29 Cortica Ltd. Configuring spanning elements of a signature generator
US11750639B2 (en) 2021-04-05 2023-09-05 Bank Of America Corporation ATM-based anomaly and security threat detection
US11758004B2 (en) 2005-10-26 2023-09-12 Cortica Ltd. System and method for providing recommendations based on user profiles
US11760387B2 (en) 2017-07-05 2023-09-19 AutoBrains Technologies Ltd. Driving policies determination
US11827215B2 (en) 2020-03-31 2023-11-28 AutoBrains Technologies Ltd. Method for training a driving related object detector
US11899707B2 (en) 2017-07-09 2024-02-13 Cortica Ltd. Driving policies determination
US11908242B2 (en) 2019-03-31 2024-02-20 Cortica Ltd. Efficient calculation of a robust signature of a media unit
US11904863B2 (en) 2018-10-26 2024-02-20 AutoBrains Technologies Ltd. Passing a curve
US11922293B2 (en) 2005-10-26 2024-03-05 Cortica Ltd. Computing device, a system and a method for parallel processing of data streams
US11954168B2 (en) 2020-03-31 2024-04-09 Cortica Ltd. System and method thereof for dynamically associating a link to an information resource with a multimedia content displayed in a web-page

Families Citing this family (323)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5666157A (en) 1995-01-03 1997-09-09 Arc Incorporated Abnormality detection and surveillance system
US6028626A (en) 1995-01-03 2000-02-22 Arc Incorporated Abnormality detection and surveillance system
CA2290383A1 (en) * 1997-06-04 1998-12-10 Ascom Systec Ag Method for surveying a predetermined surveillance area
US6172605B1 (en) * 1997-07-02 2001-01-09 Matsushita Electric Industrial Co., Ltd. Remote monitoring system and method
JP3192663B2 (en) * 1997-07-11 2001-07-30 三菱電機株式会社 Video collection device
US7088387B1 (en) * 1997-08-05 2006-08-08 Mitsubishi Electric Research Laboratories, Inc. Video recording device responsive to triggering event
US6508397B1 (en) * 1998-03-30 2003-01-21 Citicorp Development Center, Inc. Self-defense ATM
JP3566546B2 (en) * 1998-04-01 2004-09-15 Kddi株式会社 Image quality abnormality detection method and apparatus
US6466258B1 (en) * 1999-02-12 2002-10-15 Lockheed Martin Corporation 911 real time information communication
GB2351585B (en) * 1999-06-29 2003-09-03 Ncr Int Inc Self service terminal
US7005985B1 (en) * 1999-07-20 2006-02-28 Axcess, Inc. Radio frequency identification system and method
GB9918248D0 (en) 1999-08-04 1999-10-06 Matra Bae Dynamics Uk Ltd Improvements in and relating to surveillance systems
GB2360863B (en) * 1999-11-15 2002-08-28 Scott C Harris Automatic cell phone detection at a combustible delivery station
US7286158B1 (en) * 1999-12-22 2007-10-23 Axcess International Inc. Method and system for providing integrated remote monitoring services
US6940998B2 (en) * 2000-02-04 2005-09-06 Cernium, Inc. System for automated screening of security cameras
US6804232B1 (en) * 2000-03-27 2004-10-12 Bbnt Solutions Llc Personal area network with automatic attachment and detachment
US7106887B2 (en) * 2000-04-13 2006-09-12 Fuji Photo Film Co., Ltd. Image processing method using conditions corresponding to an identified person
DE60011945T2 (en) * 2000-04-24 2005-08-25 Video Domain Technologies Ltd. MONITORING SYSTEM WITH CAMERA
US6701005B1 (en) 2000-04-29 2004-03-02 Cognex Corporation Method and apparatus for three-dimensional object segmentation
US7433493B1 (en) 2000-09-06 2008-10-07 Hitachi, Ltd. Abnormal behavior detector
US8564661B2 (en) 2000-10-24 2013-10-22 Objectvideo, Inc. Video analytic rule detection system and method
US20050162515A1 (en) * 2000-10-24 2005-07-28 Objectvideo, Inc. Video surveillance system
US8711217B2 (en) * 2000-10-24 2014-04-29 Objectvideo, Inc. Video surveillance system employing video primitives
GB2368482B (en) * 2000-10-26 2004-08-25 Hewlett Packard Co Optimal image capture
WO2002037856A1 (en) * 2000-11-06 2002-05-10 Dynapel Systems, Inc. Surveillance video camera enhancement system
WO2002045434A1 (en) * 2000-12-01 2002-06-06 Vigilos, Inc. System and method for processing video data utilizing motion detection and subdivided video fields
US7307999B1 (en) 2001-02-16 2007-12-11 Bbn Technologies Corp. Systems and methods that identify normal traffic during network attacks
JP3926572B2 (en) * 2001-03-02 2007-06-06 株式会社日立製作所 Image monitoring method, image monitoring apparatus, and storage medium
US20090231436A1 (en) * 2001-04-19 2009-09-17 Faltesek Anthony E Method and apparatus for tracking with identification
GB0110480D0 (en) * 2001-04-28 2001-06-20 Univ Manchester Metropolitan Methods and apparatus for analysing the behaviour of a subject
US20020171734A1 (en) * 2001-05-16 2002-11-21 Hiroshi Arakawa Remote monitoring system
AT411500B (en) * 2001-06-12 2004-01-26 Siemens Ag Oesterreich DUAL TOLL SYSTEM
US7580912B2 (en) * 2001-06-12 2009-08-25 Alcatel-Lucent Usa Inc. Performance data mining based on real time analysis of sensor data
US7953219B2 (en) * 2001-07-19 2011-05-31 Nice Systems, Ltd. Method apparatus and system for capturing and analyzing interaction based content
GB2378301A (en) * 2001-07-31 2003-02-05 Hewlett Packard Co Personal object recognition system for visually impaired persons
US7728870B2 (en) 2001-09-06 2010-06-01 Nice Systems Ltd Advanced quality management and recording solutions for walk-in environments
EP1423967A2 (en) * 2001-09-06 2004-06-02 Nice Systems Ltd. Recording of interactions between a customer and a sales person at a point of sales
NZ514311A (en) * 2001-09-20 2004-03-26 Smart Surveillance Technologie Intelligent video surveillance system for a service station forecourt
AU2002337592A1 (en) * 2001-09-24 2003-04-07 Nice Systems Ltd. System and method for the automatic control of video frame rate
US20030095180A1 (en) * 2001-11-21 2003-05-22 Montgomery Dennis L. Method and system for size adaptation and storage minimization source noise correction, and source watermarking of digital data frames
US7688349B2 (en) * 2001-12-07 2010-03-30 International Business Machines Corporation Method of detecting and tracking groups of people
US20030107650A1 (en) * 2001-12-11 2003-06-12 Koninklijke Philips Electronics N.V. Surveillance system with suspicious behavior detection
JP3996428B2 (en) * 2001-12-25 2007-10-24 松下電器産業株式会社 Abnormality detection device and abnormality detection system
KR20030061047A (en) * 2002-01-07 2003-07-18 이철희 Face recognition algorithm
US20050128304A1 (en) * 2002-02-06 2005-06-16 Manasseh Frederick M. System and method for traveler interactions management
US7683929B2 (en) * 2002-02-06 2010-03-23 Nice Systems, Ltd. System and method for video content analysis-based detection, surveillance and alarm management
US7436887B2 (en) 2002-02-06 2008-10-14 Playtex Products, Inc. Method and apparatus for video frame sequence-based object tracking
US7761544B2 (en) 2002-03-07 2010-07-20 Nice Systems, Ltd. Method and apparatus for internal and external monitoring of a transportation vehicle
AU2002950805A0 (en) * 2002-08-15 2002-09-12 Momentum Technologies Group Improvements relating to video transmission systems
AU2003250593B2 (en) * 2002-08-15 2008-07-17 Momentum Technologies Solutions Pty Ltd Method and computer program for providing visual information to a viewer
JP2005537608A (en) * 2002-09-02 2005-12-08 サムスン エレクトロニクス カンパニー リミテッド Optical information storage medium, method and apparatus for recording and / or reproducing information on and / or from optical information storage medium
JP2004096518A (en) * 2002-09-02 2004-03-25 Japan Servo Co Ltd Monitoring camera driving method using rotary electric machine
US7002472B2 (en) * 2002-09-04 2006-02-21 Northrop Grumman Corporation Smart and secure container
US7400744B2 (en) * 2002-09-05 2008-07-15 Cognex Technology And Investment Corporation Stereo door sensor
US7397929B2 (en) * 2002-09-05 2008-07-08 Cognex Technology And Investment Corporation Method and apparatus for monitoring a passageway using 3D images
US7920718B2 (en) * 2002-09-05 2011-04-05 Cognex Corporation Multi-zone passageway monitoring system and method
US7152051B1 (en) * 2002-09-30 2006-12-19 Michael Lamport Commons Intelligent control with hierarchical stacked neural networks
US8154581B2 (en) 2002-10-15 2012-04-10 Revolutionary Concepts, Inc. Audio-video communication system for receiving person at entrance
US6862253B2 (en) * 2002-10-23 2005-03-01 Robert L. Blosser Sonic identification system and method
US7221775B2 (en) * 2002-11-12 2007-05-22 Intellivid Corporation Method and apparatus for computerized image background analysis
CA2505831C (en) * 2002-11-12 2014-06-10 Intellivid Corporation Method and system for tracking and behavioral monitoring of multiple objects moving through multiple fields-of-view
JP2004172780A (en) * 2002-11-19 2004-06-17 Hitachi Ltd Camera system, camera apparatus, and recording apparatus
US6987451B2 (en) * 2002-12-03 2006-01-17 3Rd Millennium Solutions. Ltd. Surveillance system with identification correlation
US6791603B2 (en) 2002-12-03 2004-09-14 Sensormatic Electronics Corporation Event driven video tracking system
US7400344B2 (en) * 2002-12-19 2008-07-15 Hitachi Kokusai Electric Inc. Object tracking method and object tracking apparatus
US7151454B2 (en) * 2003-01-02 2006-12-19 Covi Technologies Systems and methods for location of objects
US20040148518A1 (en) * 2003-01-27 2004-07-29 John Grundback Distributed surveillance system
US8614741B2 (en) * 2003-03-31 2013-12-24 Alcatel Lucent Method and apparatus for intelligent and automatic sensor control using multimedia database system
EP1616309A1 (en) * 2003-04-09 2006-01-18 Nice Systems Ltd. Apparatus, system and method for dispute resolution, regulation compliance and quality management in financial institutions
US7412392B1 (en) * 2003-04-14 2008-08-12 Sprint Communications Company L.P. Conference multi-tasking system and method
US20040212678A1 (en) * 2003-04-25 2004-10-28 Cooper Peter David Low power motion detection system
US7643055B2 (en) * 2003-04-25 2010-01-05 Aptina Imaging Corporation Motion detecting camera system
US20040223054A1 (en) * 2003-05-06 2004-11-11 Rotholtz Ben Aaron Multi-purpose video surveillance
JP3829829B2 (en) * 2003-08-06 2006-10-04 コニカミノルタホールディングス株式会社 Control device, program, and control method
US7546173B2 (en) * 2003-08-18 2009-06-09 Nice Systems, Ltd. Apparatus and method for audio content analysis, marking and summing
US7295106B1 (en) * 2003-09-03 2007-11-13 Siemens Schweiz Ag Systems and methods for classifying objects within a monitored zone using multiple surveillance devices
US7286157B2 (en) * 2003-09-11 2007-10-23 Intellivid Corporation Computerized method and apparatus for determining field-of-view relationships among multiple image sensors
US7049965B2 (en) 2003-10-02 2006-05-23 General Electric Company Surveillance systems and methods
US7280673B2 (en) * 2003-10-10 2007-10-09 Intellivid Corporation System and method for searching for changes in surveillance video
US7346187B2 (en) * 2003-10-10 2008-03-18 Intellivid Corporation Method of counting objects in a monitored environment and apparatus for the same
US20050086698A1 (en) * 2003-10-20 2005-04-21 Wang Cheng-Yu Automatic monitoring and alerting device
US7623674B2 (en) * 2003-11-05 2009-11-24 Cognex Technology And Investment Corporation Method and system for enhanced portal security through stereoscopy
WO2005046195A1 (en) * 2003-11-05 2005-05-19 Nice Systems Ltd. Apparatus and method for event-driven content analysis
US8326084B1 (en) 2003-11-05 2012-12-04 Cognex Technology And Investment Corporation System and method of auto-exposure control for image acquisition hardware using three dimensional information
US7148912B2 (en) * 2003-11-17 2006-12-12 Vidient Systems, Inc. Video surveillance system in which trajectory hypothesis spawning allows for trajectory splitting and/or merging
US7136507B2 (en) * 2003-11-17 2006-11-14 Vidient Systems, Inc. Video surveillance system with rule-based reasoning and multiple-hypothesis scoring
US7088846B2 (en) * 2003-11-17 2006-08-08 Vidient Systems, Inc. Video surveillance system that detects predefined behaviors based on predetermined patterns of movement through zones
US7127083B2 (en) * 2003-11-17 2006-10-24 Vidient Systems, Inc. Video surveillance system with object detection and probability scoring based on object class
US7664292B2 (en) * 2003-12-03 2010-02-16 Safehouse International, Inc. Monitoring an output from a camera
EP1634225A4 (en) * 2004-03-10 2008-01-16 Nice Systems Ltd Apparatus and method for generating a content-based follow up
US20050232580A1 (en) * 2004-03-11 2005-10-20 Interdigital Technology Corporation Control of device operation within an area
US7841120B2 (en) 2004-03-22 2010-11-30 Wilcox Industries Corp. Hand grip apparatus for firearm
US7086139B2 (en) * 2004-04-30 2006-08-08 Hitachi Global Storage Technologies Netherlands B.V. Methods of making magnetic write heads using electron beam lithography
US8204884B2 (en) * 2004-07-14 2012-06-19 Nice Systems Ltd. Method, apparatus and system for capturing and analyzing interaction based content
WO2006021943A1 (en) * 2004-08-09 2006-03-02 Nice Systems Ltd. Apparatus and method for multimedia content based
US8724891B2 (en) * 2004-08-31 2014-05-13 Ramot At Tel-Aviv University Ltd. Apparatus and methods for the detection of abnormal motion in a video stream
US8078463B2 (en) * 2004-11-23 2011-12-13 Nice Systems, Ltd. Method and apparatus for speaker spotting
US20060137018A1 (en) * 2004-11-29 2006-06-22 Interdigital Technology Corporation Method and apparatus to provide secured surveillance data to authorized entities
US20060227640A1 (en) * 2004-12-06 2006-10-12 Interdigital Technology Corporation Sensing device with activation and sensing alert functions
US7574220B2 (en) * 2004-12-06 2009-08-11 Interdigital Technology Corporation Method and apparatus for alerting a target that it is subject to sensing and restricting access to sensed content associated with the target
TWI285742B (en) * 2004-12-06 2007-08-21 Interdigital Tech Corp Method and apparatus for detecting portable electronic device functionality
EP1815365A4 (en) * 2005-02-07 2009-04-01 Nice Systems Ltd Upgrading performance using aggregated information shared between management systems
US7837013B2 (en) * 2005-03-02 2010-11-23 Mitsubishi Electric Corporation Image monitoring device for elevator with image recording
US7643056B2 (en) * 2005-03-14 2010-01-05 Aptina Imaging Corporation Motion detecting camera system
US8005675B2 (en) * 2005-03-17 2011-08-23 Nice Systems, Ltd. Apparatus and method for audio analysis
US7286056B2 (en) * 2005-03-22 2007-10-23 Lawrence Kates System and method for pest detection
US8174572B2 (en) * 2005-03-25 2012-05-08 Sensormatic Electronics, LLC Intelligent camera selection and object tracking
EP1867167A4 (en) * 2005-04-03 2009-05-06 Nice Systems Ltd Apparatus and methods for the semi-automatic tracking and examining of an object or an event in a monitored site
US7386105B2 (en) * 2005-05-27 2008-06-10 Nice Systems Ltd Method and apparatus for fraud detection
WO2007017853A1 (en) * 2005-08-08 2007-02-15 Nice Systems Ltd. Apparatus and methods for the detection of emotions in audio interactions
US9036028B2 (en) 2005-09-02 2015-05-19 Sensormatic Electronics, LLC Object tracking and alerts
CA2624657A1 (en) * 2005-10-05 2007-04-19 Redxdefense, Llc Visitor control and tracking system
US8111904B2 (en) 2005-10-07 2012-02-07 Cognex Technology And Investment Corp. Methods and apparatus for practical 3D vision system
US7738008B1 (en) * 2005-11-07 2010-06-15 Infrared Systems International, Inc. Infrared security system and method
TWI289407B (en) * 2005-11-25 2007-11-01 Benq Corp Audio adjusting system
WO2007086042A2 (en) * 2006-01-25 2007-08-02 Nice Systems Ltd. Method and apparatus for segmentation of audio interactions
JP4442571B2 (en) * 2006-02-10 2010-03-31 ソニー株式会社 Imaging apparatus and control method thereof
JP4890880B2 (en) * 2006-02-16 2012-03-07 キヤノン株式会社 Image transmitting apparatus, image transmitting method, program, and storage medium
US20070205896A1 (en) * 2006-03-02 2007-09-06 Axcess International Inc. System and Method for Determining Location, Directionality, and Velocity of RFID Tags
ITRM20060136A1 (en) * 2006-03-10 2007-09-11 Link Formazione S R L INTERACTIVE MULTIMEDIA SYSTEM
US7978698B2 (en) * 2006-03-16 2011-07-12 Panasonic Corporation Terminal for performing multiple access transmission suitable to a transmission path having varied characteristics
WO2007109241A2 (en) * 2006-03-20 2007-09-27 Axcess International Inc. Multi-tag tracking systems and methods
US8848053B2 (en) * 2006-03-28 2014-09-30 Objectvideo, Inc. Automatic extraction of secondary video streams
US20070250898A1 (en) * 2006-03-28 2007-10-25 Object Video, Inc. Automatic extraction of secondary video streams
WO2007118272A1 (en) * 2006-04-13 2007-10-25 Curtin University Of Technology Virtual observer
US8725518B2 (en) * 2006-04-25 2014-05-13 Nice Systems Ltd. Automatic speech analysis
US20070252693A1 (en) * 2006-05-01 2007-11-01 Jocelyn Janson System and method for surveilling a scene
US7800503B2 (en) * 2006-05-11 2010-09-21 Axcess International Inc. Radio frequency identification (RFID) tag antenna design
US7770221B2 (en) * 2006-05-18 2010-08-03 Nice Systems, Ltd. Method and apparatus for combining traffic analysis and monitoring center in lawful interception
US20070281760A1 (en) * 2006-05-23 2007-12-06 Intermec Ip Corp. Wireless, batteryless, audio communications device
US7825792B2 (en) * 2006-06-02 2010-11-02 Sensormatic Electronics Llc Systems and methods for distributed monitoring of remote sites
US7671728B2 (en) 2006-06-02 2010-03-02 Sensormatic Electronics, LLC Systems and methods for distributed monitoring of remote sites
US8009193B2 (en) * 2006-06-05 2011-08-30 Fuji Xerox Co., Ltd. Unusual event detection via collaborative video mining
US20070290881A1 (en) * 2006-06-13 2007-12-20 Intermec Ip Corp. Wireless remote control, system and method
US7411497B2 (en) * 2006-08-15 2008-08-12 Lawrence Kates System and method for intruder detection
US7822605B2 (en) * 2006-10-19 2010-10-26 Nice Systems Ltd. Method and apparatus for large population speaker identification in telephone interactions
US7631046B2 (en) * 2006-10-26 2009-12-08 Nice Systems, Ltd. Method and apparatus for lawful interception of web based messaging communication
US7577246B2 (en) * 2006-12-20 2009-08-18 Nice Systems Ltd. Method and system for automatic quality evaluation
US8295542B2 (en) * 2007-01-12 2012-10-23 International Business Machines Corporation Adjusting a consumer experience based on a 3D captured image stream of a consumer response
US8269834B2 (en) 2007-01-12 2012-09-18 International Business Machines Corporation Warning a user about adverse behaviors of others within an environment based on a 3D captured image stream
US8588464B2 (en) * 2007-01-12 2013-11-19 International Business Machines Corporation Assisting a vision-impaired user with navigation based on a 3D captured image stream
US20080189171A1 (en) * 2007-02-01 2008-08-07 Nice Systems Ltd. Method and apparatus for call categorization
US8571853B2 (en) * 2007-02-11 2013-10-29 Nice Systems Ltd. Method and system for laughter detection
JP5121258B2 (en) * 2007-03-06 2013-01-16 株式会社東芝 Suspicious behavior detection system and method
US7599475B2 (en) * 2007-03-12 2009-10-06 Nice Systems, Ltd. Method and apparatus for generic analytics
US20080243439A1 (en) * 2007-03-28 2008-10-02 Runkle Paul R Sensor exploration and management through adaptive sensing framework
US20080243425A1 (en) * 2007-03-28 2008-10-02 Eliazar Austin I D Tracking target objects through occlusions
US9361623B2 (en) * 2007-04-03 2016-06-07 International Business Machines Corporation Preferred customer marketing delivery based on biometric data for a customer
US9031857B2 (en) * 2007-04-03 2015-05-12 International Business Machines Corporation Generating customized marketing messages at the customer level based on biometric data
US20080249864A1 (en) * 2007-04-03 2008-10-09 Robert Lee Angell Generating customized marketing content to improve cross sale of related items
US8639563B2 (en) * 2007-04-03 2014-01-28 International Business Machines Corporation Generating customized marketing messages at a customer level using current events data
US8775238B2 (en) * 2007-04-03 2014-07-08 International Business Machines Corporation Generating customized disincentive marketing content for a customer based on customer risk assessment
US20080249870A1 (en) * 2007-04-03 2008-10-09 Robert Lee Angell Method and apparatus for decision tree based marketing and selling for a retail store
US9846883B2 (en) * 2007-04-03 2017-12-19 International Business Machines Corporation Generating customized marketing messages using automatically generated customer identification data
US20080249866A1 (en) * 2007-04-03 2008-10-09 Robert Lee Angell Generating customized marketing content for upsale of items
US9092808B2 (en) * 2007-04-03 2015-07-28 International Business Machines Corporation Preferred customer marketing delivery based on dynamic data for a customer
US20080249865A1 (en) * 2007-04-03 2008-10-09 Robert Lee Angell Recipe and project based marketing and guided selling in a retail store environment
US8812355B2 (en) * 2007-04-03 2014-08-19 International Business Machines Corporation Generating customized marketing messages for a customer using dynamic customer behavior data
US8831972B2 (en) * 2007-04-03 2014-09-09 International Business Machines Corporation Generating a customer risk assessment using dynamic customer data
US9685048B2 (en) * 2007-04-03 2017-06-20 International Business Machines Corporation Automatically generating an optimal marketing strategy for improving cross sales and upsales of items
US20080249858A1 (en) * 2007-04-03 2008-10-09 Robert Lee Angell Automatically generating an optimal marketing model for marketing products to customers
US9031858B2 (en) * 2007-04-03 2015-05-12 International Business Machines Corporation Using biometric data for a customer to improve upsale ad cross-sale of items
US9626684B2 (en) * 2007-04-03 2017-04-18 International Business Machines Corporation Providing customized digital media marketing content directly to a customer
US20080249835A1 (en) * 2007-04-03 2008-10-09 Robert Lee Angell Identifying significant groupings of customers for use in customizing digital media marketing content provided directly to a customer
US8126260B2 (en) * 2007-05-29 2012-02-28 Cognex Corporation System and method for locating a three-dimensional object using machine vision
JP2010533319A (en) * 2007-06-09 2010-10-21 センサーマティック・エレクトロニクス・コーポレーション Systems and methods for integrating video analysis and data analysis / mining
US20090005650A1 (en) * 2007-06-29 2009-01-01 Robert Lee Angell Method and apparatus for implementing digital video modeling to generate a patient risk assessment model
US7908237B2 (en) * 2007-06-29 2011-03-15 International Business Machines Corporation Method and apparatus for identifying unexpected behavior of a customer in a retail environment using detected location data, temperature, humidity, lighting conditions, music, and odors
US20090006125A1 (en) * 2007-06-29 2009-01-01 Robert Lee Angell Method and apparatus for implementing digital video modeling to generate an optimal healthcare delivery model
US7908233B2 (en) * 2007-06-29 2011-03-15 International Business Machines Corporation Method and apparatus for implementing digital video modeling to generate an expected behavior model
US20090012826A1 (en) * 2007-07-02 2009-01-08 Nice Systems Ltd. Method and apparatus for adaptive interaction analytics
US20090083121A1 (en) * 2007-09-26 2009-03-26 Robert Lee Angell Method and apparatus for determining profitability of customer groups identified from a continuous video stream
US8195499B2 (en) * 2007-09-26 2012-06-05 International Business Machines Corporation Identifying customer behavioral types from a continuous video stream for use in optimizing loss leader merchandizing
US20090089107A1 (en) * 2007-09-27 2009-04-02 Robert Lee Angell Method and apparatus for ranking a customer using dynamically generated external data
US20090122144A1 (en) * 2007-11-14 2009-05-14 Joel Pat Latham Method for detecting events at a secured location
MY143022A (en) * 2007-11-23 2011-02-14 Mimos Berhad Method for detecting unattended object and removal of static object
US20090150321A1 (en) * 2007-12-07 2009-06-11 Nokia Corporation Method, Apparatus and Computer Program Product for Developing and Utilizing User Pattern Profiles
JP5141317B2 (en) * 2008-03-14 2013-02-13 オムロン株式会社 Target image detection device, control program, recording medium storing the program, and electronic apparatus including the target image detection device
JP4525795B2 (en) * 2008-05-16 2010-08-18 ソニー株式会社 Reception device, reception method, program, and communication system
SI22822A (en) 2008-06-23 2009-12-31 Institut JoĹľef Stefan System and device for speech-controlled telephone communication
US8638194B2 (en) * 2008-07-25 2014-01-28 Axcess International, Inc. Multiple radio frequency identification (RFID) tag wireless wide area network (WWAN) protocol
US9470704B2 (en) 2009-02-23 2016-10-18 Nortek Security & Control Llc Wearable motion sensing device
US8560267B2 (en) 2009-09-15 2013-10-15 Imetrikus, Inc. Identifying one or more activities of an animate or inanimate object
US8972197B2 (en) * 2009-09-15 2015-03-03 Numera, Inc. Method and system for analyzing breathing of a user
US8571261B2 (en) 2009-04-22 2013-10-29 Checkvideo Llc System and method for motion detection in a surveillance video
US8340352B2 (en) * 2009-08-18 2012-12-25 Behavioral Recognition Systems, Inc. Inter-trajectory anomaly detection using adaptive voting experts in a video surveillance system
JP5570176B2 (en) * 2009-10-19 2014-08-13 キヤノン株式会社 Image processing system and information processing method
US20110109747A1 (en) * 2009-11-12 2011-05-12 Siemens Industry, Inc. System and method for annotating video with geospatially referenced data
US20120076368A1 (en) * 2010-09-27 2012-03-29 David Staudacher Face identification based on facial feature changes
TW201216713A (en) * 2010-10-13 2012-04-16 Hon Hai Prec Ind Co Ltd System and method for automatically adjusting camera resolutions
US8775341B1 (en) 2010-10-26 2014-07-08 Michael Lamport Commons Intelligent control with hierarchical stacked neural networks
US9015093B1 (en) 2010-10-26 2015-04-21 Michael Lamport Commons Intelligent control with hierarchical stacked neural networks
KR20120071553A (en) * 2010-12-23 2012-07-03 한국전자통신연구원 Apparatus and method for detection of threat
US8959042B1 (en) 2011-04-18 2015-02-17 The Boeing Company Methods and systems for estimating subject cost from surveillance
IL212972A0 (en) * 2011-05-18 2011-07-31 Zlotnikov Andrey Counter surveillance system
JP5917270B2 (en) * 2011-05-27 2016-05-11 キヤノン株式会社 Sound detection apparatus, control method therefor, and program
CN103931172A (en) * 2011-06-10 2014-07-16 菲力尔系统公司 Systems and methods for intelligent monitoring of thoroughfares using thermal imaging
US8942994B2 (en) * 2012-02-09 2015-01-27 Aic Innovations Group, Inc. Method and apparatus for encouraging consumption of a product
KR20130136251A (en) * 2012-06-04 2013-12-12 한국전자통신연구원 Method and apparatus for situation recognition using object energy function
US9202520B1 (en) * 2012-10-17 2015-12-01 Amazon Technologies, Inc. Systems and methods for determining content preferences based on vocal utterances and/or movement by a user
TWI486915B (en) * 2012-10-25 2015-06-01 Hon Hai Prec Ind Co Ltd Personal protective equipment, danger early-waning system and method
CA2897910C (en) * 2013-01-29 2018-07-17 Ramrock Video Technology Laboratory Co., Ltd. Surveillance system
US9483732B1 (en) 2013-02-08 2016-11-01 Marko Milakovich High value information alert and reporting system and method
WO2014144841A1 (en) * 2013-03-15 2014-09-18 E-Connect Visual analysis of transactions
US10437658B2 (en) 2013-06-06 2019-10-08 Zebra Technologies Corporation Method, apparatus, and computer program product for collecting and displaying sporting event data based on real time data for proximity and movement of objects
US10609762B2 (en) 2013-06-06 2020-03-31 Zebra Technologies Corporation Method, apparatus, and computer program product improving backhaul of sensor and other data to real time location system network
US11423464B2 (en) 2013-06-06 2022-08-23 Zebra Technologies Corporation Method, apparatus, and computer program product for enhancement of fan experience based on location data
US9699278B2 (en) 2013-06-06 2017-07-04 Zih Corp. Modular location tag for a real time location system network
US9517417B2 (en) 2013-06-06 2016-12-13 Zih Corp. Method, apparatus, and computer program product for performance analytics determining participant statistical data and game status data
US9715005B2 (en) 2013-06-06 2017-07-25 Zih Corp. Method, apparatus, and computer program product improving real time location systems with multiple location technologies
US9180357B2 (en) 2013-06-06 2015-11-10 Zih Corp. Multiple antenna interference rejection in ultra-wideband real time locating systems
US9342936B2 (en) 2013-07-26 2016-05-17 SkyBell Technologies, Inc. Smart lock systems and methods
US9142214B2 (en) 2013-07-26 2015-09-22 SkyBell Technologies, Inc. Light socket cameras
US9094584B2 (en) 2013-07-26 2015-07-28 SkyBell Technologies, Inc. Doorbell communication systems and methods
US11889009B2 (en) 2013-07-26 2024-01-30 Skybell Technologies Ip, Llc Doorbell communication and electrical systems
US11764990B2 (en) 2013-07-26 2023-09-19 Skybell Technologies Ip, Llc Doorbell communications systems and methods
US10708404B2 (en) 2014-09-01 2020-07-07 Skybell Technologies Ip, Llc Doorbell communication and electrical systems
US11651665B2 (en) 2013-07-26 2023-05-16 Skybell Technologies Ip, Llc Doorbell communities
US9118819B1 (en) 2013-07-26 2015-08-25 SkyBell Technologies, Inc. Doorbell communication systems and methods
US10440165B2 (en) 2013-07-26 2019-10-08 SkyBell Technologies, Inc. Doorbell communication and electrical systems
US9058738B1 (en) 2013-07-26 2015-06-16 SkyBell Technologies, Inc. Doorbell communication systems and methods
US10672238B2 (en) 2015-06-23 2020-06-02 SkyBell Technologies, Inc. Doorbell communities
US9113051B1 (en) 2013-07-26 2015-08-18 SkyBell Technologies, Inc. Power outlet cameras
US10204467B2 (en) 2013-07-26 2019-02-12 SkyBell Technologies, Inc. Smart lock systems and methods
US9736284B2 (en) 2013-07-26 2017-08-15 SkyBell Technologies, Inc. Doorbell communication and electrical systems
US10044519B2 (en) 2015-01-05 2018-08-07 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9769435B2 (en) 2014-08-11 2017-09-19 SkyBell Technologies, Inc. Monitoring systems and methods
US9113052B1 (en) 2013-07-26 2015-08-18 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9172922B1 (en) 2013-12-06 2015-10-27 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9060103B2 (en) 2013-07-26 2015-06-16 SkyBell Technologies, Inc. Doorbell security and safety
US9179108B1 (en) 2013-07-26 2015-11-03 SkyBell Technologies, Inc. Doorbell chime systems and methods
US9065987B2 (en) 2013-07-26 2015-06-23 SkyBell Technologies, Inc. Doorbell communication systems and methods
US11909549B2 (en) 2013-07-26 2024-02-20 Skybell Technologies Ip, Llc Doorbell communication systems and methods
US9247219B2 (en) 2013-07-26 2016-01-26 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9179107B1 (en) 2013-07-26 2015-11-03 SkyBell Technologies, Inc. Doorbell chime systems and methods
US9196133B2 (en) 2013-07-26 2015-11-24 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9172920B1 (en) 2014-09-01 2015-10-27 SkyBell Technologies, Inc. Doorbell diagnostics
US9013575B2 (en) * 2013-07-26 2015-04-21 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9060104B2 (en) 2013-07-26 2015-06-16 SkyBell Technologies, Inc. Doorbell communication systems and methods
US10733823B2 (en) 2013-07-26 2020-08-04 Skybell Technologies Ip, Llc Garage door communication systems and methods
US11004312B2 (en) 2015-06-23 2021-05-11 Skybell Technologies Ip, Llc Doorbell communities
US9179109B1 (en) 2013-12-06 2015-11-03 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9049352B2 (en) 2013-07-26 2015-06-02 SkyBell Technologies, Inc. Pool monitor systems and methods
US20170263067A1 (en) 2014-08-27 2017-09-14 SkyBell Technologies, Inc. Smart lock systems and methods
US9172921B1 (en) 2013-12-06 2015-10-27 SkyBell Technologies, Inc. Doorbell antenna
US9230424B1 (en) 2013-12-06 2016-01-05 SkyBell Technologies, Inc. Doorbell communities
US9237318B2 (en) 2013-07-26 2016-01-12 SkyBell Technologies, Inc. Doorbell communication systems and methods
US20180343141A1 (en) 2015-09-22 2018-11-29 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9197867B1 (en) 2013-12-06 2015-11-24 SkyBell Technologies, Inc. Identity verification using a social network
US9160987B1 (en) 2013-07-26 2015-10-13 SkyBell Technologies, Inc. Doorbell chime systems and methods
KR102066939B1 (en) * 2013-09-27 2020-01-16 한화테크윈 주식회사 Image monitoring system
US9743049B2 (en) 2013-12-06 2017-08-22 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9786133B2 (en) 2013-12-06 2017-10-10 SkyBell Technologies, Inc. Doorbell chime systems and methods
US9799183B2 (en) 2013-12-06 2017-10-24 SkyBell Technologies, Inc. Doorbell package detection systems and methods
US9253455B1 (en) 2014-06-25 2016-02-02 SkyBell Technologies, Inc. Doorbell communication systems and methods
GB2541834B (en) 2014-06-05 2020-12-23 Zebra Tech Corp Receiver processor for adaptive windowing and high-resolution TOA determination in a multiple receiver target location system
US9661455B2 (en) 2014-06-05 2017-05-23 Zih Corp. Method, apparatus, and computer program product for real time location system referencing in physically and radio frequency challenged environments
US20150375083A1 (en) 2014-06-05 2015-12-31 Zih Corp. Method, Apparatus, And Computer Program Product For Enhancement Of Event Visualizations Based On Location Data
US9626616B2 (en) 2014-06-05 2017-04-18 Zih Corp. Low-profile real-time location system tag
US9668164B2 (en) 2014-06-05 2017-05-30 Zih Corp. Receiver processor for bandwidth management of a multiple receiver real-time location system (RTLS)
GB2542298B (en) 2014-06-05 2021-01-20 Zebra Tech Corp Method for iterative target location in a multiple receiver target location system
GB2541617B (en) 2014-06-05 2021-07-07 Zebra Tech Corp Systems, apparatus and methods for variable rate ultra-wideband communications
US9759803B2 (en) 2014-06-06 2017-09-12 Zih Corp. Method, apparatus, and computer program product for employing a spatial association model in a real time location system
WO2015186043A1 (en) 2014-06-06 2015-12-10 Zih Corp. Method, apparatus, and computer program product improving real time location systems with multiple location technologies
US20170085843A1 (en) 2015-09-22 2017-03-23 SkyBell Technologies, Inc. Doorbell communication systems and methods
US11184589B2 (en) 2014-06-23 2021-11-23 Skybell Technologies Ip, Llc Doorbell communication systems and methods
US9888216B2 (en) 2015-09-22 2018-02-06 SkyBell Technologies, Inc. Doorbell communication systems and methods
US10687029B2 (en) 2015-09-22 2020-06-16 SkyBell Technologies, Inc. Doorbell communication systems and methods
US9449229B1 (en) * 2014-07-07 2016-09-20 Google Inc. Systems and methods for categorizing motion event candidates
US9501915B1 (en) * 2014-07-07 2016-11-22 Google Inc. Systems and methods for analyzing a video stream
US9224044B1 (en) 2014-07-07 2015-12-29 Google Inc. Method and system for video zone monitoring
US10140827B2 (en) 2014-07-07 2018-11-27 Google Llc Method and system for processing motion event notifications
US10127783B2 (en) 2014-07-07 2018-11-13 Google Llc Method and device for processing motion events
US9997036B2 (en) 2015-02-17 2018-06-12 SkyBell Technologies, Inc. Power outlet cameras
US9009805B1 (en) 2014-09-30 2015-04-14 Google Inc. Method and system for provisioning an electronic device
US9811989B2 (en) * 2014-09-30 2017-11-07 The Boeing Company Event detection system
USD782495S1 (en) 2014-10-07 2017-03-28 Google Inc. Display screen or portion thereof with graphical user interface
US10108325B2 (en) 2014-12-11 2018-10-23 Rdi Technologies, Inc. Method of analyzing, displaying, organizing and responding to vital signals
US10062411B2 (en) * 2014-12-11 2018-08-28 Jeffrey R. Hay Apparatus and method for visualizing periodic motions in mechanical components
US10742938B2 (en) 2015-03-07 2020-08-11 Skybell Technologies Ip, Llc Garage door communication systems and methods
US11575537B2 (en) 2015-03-27 2023-02-07 Skybell Technologies Ip, Llc Doorbell communication systems and methods
KR102360453B1 (en) 2015-04-10 2022-02-09 삼성전자 주식회사 Apparatus And Method For Setting A Camera
US11381686B2 (en) 2015-04-13 2022-07-05 Skybell Technologies Ip, Llc Power outlet cameras
US11641452B2 (en) 2015-05-08 2023-05-02 Skybell Technologies Ip, Llc Doorbell communication systems and methods
US9361011B1 (en) 2015-06-14 2016-06-07 Google Inc. Methods and systems for presenting multiple live video feeds in a user interface
US20180047269A1 (en) 2015-06-23 2018-02-15 SkyBell Technologies, Inc. Doorbell communities
US10706702B2 (en) 2015-07-30 2020-07-07 Skybell Technologies Ip, Llc Doorbell package detection systems and methods
US10139281B2 (en) 2015-12-04 2018-11-27 Amazon Technologies, Inc. Motion detection for A/V recording and communication devices
US10325625B2 (en) 2015-12-04 2019-06-18 Amazon Technologies, Inc. Motion detection for A/V recording and communication devices
WO2017130902A1 (en) * 2016-01-29 2017-08-03 日本電気株式会社 Information processing device, information processing method, and program
US20190208168A1 (en) * 2016-01-29 2019-07-04 John K. Collings, III Limited Access Community Surveillance System
US10997422B2 (en) * 2016-01-29 2021-05-04 Nec Corporation Information processing apparatus, information processing method, and program
US10506237B1 (en) 2016-05-27 2019-12-10 Google Llc Methods and devices for dynamic adaptation of encoding bitrate for video streaming
US10043332B2 (en) 2016-05-27 2018-08-07 SkyBell Technologies, Inc. Doorbell package detection systems and methods
ITUA20164265A1 (en) * 2016-06-10 2017-12-10 Masi Ivano Tommaso Ippazio De METHOD AND DYNAMIC MONITORING SYSTEM OF AN ENVIRONMENT
US10380429B2 (en) 2016-07-11 2019-08-13 Google Llc Methods and systems for person detection in a video feed
TWI616086B (en) * 2016-10-20 2018-02-21 Detection method of abnormal type of photographing device screen
US10691950B2 (en) 2017-03-10 2020-06-23 Turing Video, Inc. Activity recognition method and system
US11475671B2 (en) 2017-05-26 2022-10-18 Turing Video Multiple robots assisted surveillance system
US11783010B2 (en) 2017-05-30 2023-10-10 Google Llc Systems and methods of person recognition in video streams
US10417484B2 (en) 2017-05-30 2019-09-17 Wipro Limited Method and system for determining an intent of a subject using behavioural pattern
US10909825B2 (en) 2017-09-18 2021-02-02 Skybell Technologies Ip, Llc Outdoor security systems and methods
US10664688B2 (en) 2017-09-20 2020-05-26 Google Llc Systems and methods of detecting and responding to a visitor to a smart home environment
US10388132B2 (en) 2017-11-17 2019-08-20 Motorola Solutions, Inc. Systems and methods for surveillance-assisted patrol
US10750133B2 (en) 2017-12-06 2020-08-18 Honeywell International Inc. Systems and methods for automatic video recording
US10832043B2 (en) * 2018-02-06 2020-11-10 Disney Enterprises, Inc. Variable resolution recognition
US10834365B2 (en) 2018-02-08 2020-11-10 Nortek Security & Control Llc Audio-visual monitoring using a virtual assistant
US11295139B2 (en) 2018-02-19 2022-04-05 Intellivision Technologies Corp. Human presence detection in edge devices
US11615623B2 (en) 2018-02-19 2023-03-28 Nortek Security & Control Llc Object detection in edge devices for barrier operation and parcel delivery
US10978050B2 (en) 2018-02-20 2021-04-13 Intellivision Technologies Corp. Audio type detection
EP3534294B1 (en) 2018-03-02 2023-09-06 Akademia Gorniczo-Hutnicza im. Stanislawa Staszica w Krakowie Computer implemented method for processing a frame sequence, digital frame processing device computer program product and monitoring system
US11423551B1 (en) 2018-10-17 2022-08-23 Rdi Technologies, Inc. Enhanced presentation methods for visualizing motion of physical structures and machinery
US10769896B1 (en) * 2019-05-01 2020-09-08 Capital One Services, Llc Counter-fraud measures for an ATM device
US11074790B2 (en) 2019-08-24 2021-07-27 Skybell Technologies Ip, Llc Doorbell communication systems and methods
CN110488258B (en) * 2019-08-30 2022-08-05 成都纳雷科技有限公司 False alarm suppression method and device in perimeter area security
US11373317B1 (en) 2020-01-24 2022-06-28 Rdi Technologies, Inc. Measuring the speed of rotation or reciprocation of a mechanical component using one or more cameras
EP3876210B1 (en) * 2020-03-02 2023-09-13 ABB Schweiz AG A low or medium switchgear monitoring system
JP7440332B2 (en) 2020-04-21 2024-02-28 株式会社日立製作所 Event analysis system and method
US11138439B1 (en) 2020-04-30 2021-10-05 Honeywell International Inc. Systems and methods for detecting patterns within video content
EP3926544B1 (en) * 2020-06-18 2024-03-13 Tata Consultancy Services Limited System and method of gesture recognition using a reservoir based convolutional spiking neural network
US11282213B1 (en) 2020-06-24 2022-03-22 Rdi Technologies, Inc. Enhanced analysis techniques using composite frequency spectrum data
US11322182B1 (en) 2020-09-28 2022-05-03 Rdi Technologies, Inc. Enhanced visualization techniques using reconstructed time waveforms
CN114091844B (en) * 2021-11-01 2023-06-02 山东心法科技有限公司 Early warning method, device and storage medium for re-crime of violent personnel

Citations (55)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4337482A (en) 1979-10-17 1982-06-29 Coutta John M Surveillance system
JPS58109373A (en) 1981-12-24 1983-06-29 フジテツク株式会社 Detector for abnormality in elevator cage
EP0183106A2 (en) 1984-11-10 1986-06-04 Matsushita Electric Works, Ltd. Visual image sensor system
JPS61260391A (en) 1985-05-14 1986-11-18 三菱電機株式会社 Monitor/controller
JPS62136988A (en) 1985-12-10 1987-06-19 Matsushita Electric Works Ltd Invasion monitoring device
US4692806A (en) 1985-07-25 1987-09-08 Rca Corporation Image-data reduction technique
JPS62222390A (en) 1986-03-24 1987-09-30 松下電工株式会社 Image recognition type wide range monitoring system
US4737847A (en) 1985-10-11 1988-04-12 Matsushita Electric Works, Ltd. Abnormality supervising system
JPH01244598A (en) 1988-03-25 1989-09-28 Toshiba Corp Picture supervisory equipment
JPH01251195A (en) 1988-03-31 1989-10-06 Toshiba Corp Monitor
DE3832353A1 (en) 1988-09-23 1990-04-05 Nolde Sylvia Expanding the function of monitoring devices in shops and department stores
JPH02151996A (en) 1988-12-03 1990-06-11 Toshiba Corp Room entry control system
GB2239369A (en) 1989-11-09 1991-06-26 Marconi Gec Ltd Image tracking
JPH0410099A (en) 1990-04-27 1992-01-14 Toshiba Corp Detector for person acting suspiciously
US5091780A (en) 1990-05-09 1992-02-25 Carnegie-Mellon University A trainable security system emthod for the same
JPH0460880A (en) 1990-06-29 1992-02-26 Shimizu Corp Moving body discrimination and analysis controlling system
US5097328A (en) 1990-10-16 1992-03-17 Boyette Robert B Apparatus and a method for sensing events from a remote location
US5099322A (en) 1990-02-27 1992-03-24 Texas Instruments Incorporated Scene change detection system and method
US5126577A (en) 1991-06-27 1992-06-30 Electro-Optical Industries, Inc. Infrared target plate handling apparatus with improved thermal control
JPH04257190A (en) 1991-02-08 1992-09-11 Toshiba Corp Moving body tracking and display device
JPH04273689A (en) 1991-02-28 1992-09-29 Hitachi Ltd Monitoring device
JPH0512578A (en) 1991-06-28 1993-01-22 Mitsubishi Electric Corp Invasion monitoring device
JPH0514892A (en) 1991-06-28 1993-01-22 Toshiba Corp Image monitor device
JPH0546771A (en) 1991-08-16 1993-02-26 Toshiba Corp Motive object detector
EP0529196A2 (en) 1991-08-29 1993-03-03 Pioneer Electronic Corporation Picture image monitoring system
WO1993005488A1 (en) 1991-09-12 1993-03-18 Electronic Data Systems Corporation Image analyser
JPH05143737A (en) 1991-11-22 1993-06-11 Ohkura Electric Co Ltd Method and device for identification with motion vector
US5243418A (en) 1990-11-27 1993-09-07 Kabushiki Kaisha Toshiba Display monitoring system for detecting and tracking an intruder in a monitor area
GB2265516A (en) 1992-03-24 1993-09-29 Sony Broadcast & Communication Motion analysis of moving images
EP0564858A2 (en) 1992-04-06 1993-10-13 Siemens Aktiengesellschaft Method for resolving clusters of moving segments
US5283644A (en) 1991-12-11 1994-02-01 Ibaraki Security Systems Co., Ltd. Crime prevention monitor system
JPH0628449A (en) 1992-07-08 1994-02-04 Matsushita Electric Ind Co Ltd Image synthesizing device
US5289275A (en) 1991-07-12 1994-02-22 Hochiki Kabushiki Kaisha Surveillance monitor system using image processing for monitoring fires and thefts
JPH06119564A (en) 1992-10-02 1994-04-28 Toshiba Corp Suspicious person detecting system
JPH06117836A (en) 1992-08-21 1994-04-28 Matsushita Electric Ind Co Ltd Image processing apparatus, controller of air conditioner, and applied equipment using the apparatus
JPH06251159A (en) 1993-03-01 1994-09-09 Nippon Telegr & Teleph Corp <Ntt> Operation recognizing device
JPH06266840A (en) 1993-03-11 1994-09-22 Hitachi Ltd Status detector for moving object
GB2277845A (en) 1993-05-03 1994-11-09 Philips Electronics Nv Monitoring system
EP0624858A1 (en) 1993-05-14 1994-11-17 Michael Josef Lantschner Method and device for detecting irregular movements of a person
US5387768A (en) 1993-09-27 1995-02-07 Otis Elevator Company Elevator passenger detector and door control system which masks portions of a hall image to determine motion and court passengers
US5396284A (en) 1993-08-20 1995-03-07 Burle Technologies, Inc. Motion detection system
US5396252A (en) 1993-09-30 1995-03-07 United Technologies Corporation Multiple target discrimination
US5416711A (en) 1993-10-18 1995-05-16 Grumman Aerospace Corporation Infra-red sensor system for intelligent vehicle highway systems
US5473364A (en) 1994-06-03 1995-12-05 David Sarnoff Research Center, Inc. Video technique for indicating moving objects from a movable platform
US5512942A (en) 1992-10-29 1996-04-30 Fujikura Ltd. Anomaly surveillance device
US5519669A (en) 1993-08-19 1996-05-21 At&T Corp. Acoustically monitored site surveillance and security system for ATM machines and other facilities
US5546072A (en) 1994-07-22 1996-08-13 Irw Inc. Alert locator
US5554983A (en) 1992-04-24 1996-09-10 Hitachi, Ltd. Object recognition system and abnormality detection system using image processing
US5555512A (en) 1993-08-19 1996-09-10 Matsushita Electric Industrial Co., Ltd. Picture processing apparatus for processing infrared pictures obtained with an infrared ray sensor and applied apparatus utilizing the picture processing apparatus
US5666157A (en) 1995-01-03 1997-09-09 Arc Incorporated Abnormality detection and surveillance system
US5747719A (en) 1997-01-21 1998-05-05 Bottesch; H. Werner Armed terrorist immobilization (ATI) system
US5809161A (en) 1992-03-20 1998-09-15 Commonwealth Scientific And Industrial Research Organisation Vehicle monitoring system
US6028626A (en) 1995-01-03 2000-02-22 Arc Incorporated Abnormality detection and surveillance system
US6050369A (en) 1994-10-07 2000-04-18 Toc Holding Company Of New York, Inc. Elevator shaftway intrusion device using optical imaging processing
US6167143A (en) 1993-05-03 2000-12-26 U.S. Philips Corporation Monitoring system

Patent Citations (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4337482A (en) 1979-10-17 1982-06-29 Coutta John M Surveillance system
JPS58109373A (en) 1981-12-24 1983-06-29 フジテツク株式会社 Detector for abnormality in elevator cage
EP0183106A2 (en) 1984-11-10 1986-06-04 Matsushita Electric Works, Ltd. Visual image sensor system
JPS61260391A (en) 1985-05-14 1986-11-18 三菱電機株式会社 Monitor/controller
US4839631A (en) 1985-05-14 1989-06-13 Mitsubishi Denki Kabushiki Kaisha Monitor control apparatus
US4692806A (en) 1985-07-25 1987-09-08 Rca Corporation Image-data reduction technique
US4737847A (en) 1985-10-11 1988-04-12 Matsushita Electric Works, Ltd. Abnormality supervising system
JPS62136988A (en) 1985-12-10 1987-06-19 Matsushita Electric Works Ltd Invasion monitoring device
JPS62222390A (en) 1986-03-24 1987-09-30 松下電工株式会社 Image recognition type wide range monitoring system
JPH01244598A (en) 1988-03-25 1989-09-28 Toshiba Corp Picture supervisory equipment
JPH01251195A (en) 1988-03-31 1989-10-06 Toshiba Corp Monitor
DE3832353A1 (en) 1988-09-23 1990-04-05 Nolde Sylvia Expanding the function of monitoring devices in shops and department stores
JPH02151996A (en) 1988-12-03 1990-06-11 Toshiba Corp Room entry control system
GB2239369A (en) 1989-11-09 1991-06-26 Marconi Gec Ltd Image tracking
US5099322A (en) 1990-02-27 1992-03-24 Texas Instruments Incorporated Scene change detection system and method
JPH0410099A (en) 1990-04-27 1992-01-14 Toshiba Corp Detector for person acting suspiciously
US5091780A (en) 1990-05-09 1992-02-25 Carnegie-Mellon University A trainable security system emthod for the same
JPH0460880A (en) 1990-06-29 1992-02-26 Shimizu Corp Moving body discrimination and analysis controlling system
US5097328A (en) 1990-10-16 1992-03-17 Boyette Robert B Apparatus and a method for sensing events from a remote location
US5243418A (en) 1990-11-27 1993-09-07 Kabushiki Kaisha Toshiba Display monitoring system for detecting and tracking an intruder in a monitor area
JPH04257190A (en) 1991-02-08 1992-09-11 Toshiba Corp Moving body tracking and display device
JPH04273689A (en) 1991-02-28 1992-09-29 Hitachi Ltd Monitoring device
US5126577A (en) 1991-06-27 1992-06-30 Electro-Optical Industries, Inc. Infrared target plate handling apparatus with improved thermal control
JPH0512578A (en) 1991-06-28 1993-01-22 Mitsubishi Electric Corp Invasion monitoring device
JPH0514892A (en) 1991-06-28 1993-01-22 Toshiba Corp Image monitor device
US5289275A (en) 1991-07-12 1994-02-22 Hochiki Kabushiki Kaisha Surveillance monitor system using image processing for monitoring fires and thefts
JPH0546771A (en) 1991-08-16 1993-02-26 Toshiba Corp Motive object detector
EP0529196A2 (en) 1991-08-29 1993-03-03 Pioneer Electronic Corporation Picture image monitoring system
WO1993005488A1 (en) 1991-09-12 1993-03-18 Electronic Data Systems Corporation Image analyser
JPH05143737A (en) 1991-11-22 1993-06-11 Ohkura Electric Co Ltd Method and device for identification with motion vector
US5283644A (en) 1991-12-11 1994-02-01 Ibaraki Security Systems Co., Ltd. Crime prevention monitor system
US5809161A (en) 1992-03-20 1998-09-15 Commonwealth Scientific And Industrial Research Organisation Vehicle monitoring system
GB2265516A (en) 1992-03-24 1993-09-29 Sony Broadcast & Communication Motion analysis of moving images
EP0564858A2 (en) 1992-04-06 1993-10-13 Siemens Aktiengesellschaft Method for resolving clusters of moving segments
US5554983A (en) 1992-04-24 1996-09-10 Hitachi, Ltd. Object recognition system and abnormality detection system using image processing
JPH0628449A (en) 1992-07-08 1994-02-04 Matsushita Electric Ind Co Ltd Image synthesizing device
JPH06117836A (en) 1992-08-21 1994-04-28 Matsushita Electric Ind Co Ltd Image processing apparatus, controller of air conditioner, and applied equipment using the apparatus
JPH06119564A (en) 1992-10-02 1994-04-28 Toshiba Corp Suspicious person detecting system
US5512942A (en) 1992-10-29 1996-04-30 Fujikura Ltd. Anomaly surveillance device
JPH06251159A (en) 1993-03-01 1994-09-09 Nippon Telegr & Teleph Corp <Ntt> Operation recognizing device
JPH06266840A (en) 1993-03-11 1994-09-22 Hitachi Ltd Status detector for moving object
DE4314483A1 (en) 1993-05-03 1994-11-10 Philips Patentverwaltung Surveillance system
US6167143A (en) 1993-05-03 2000-12-26 U.S. Philips Corporation Monitoring system
GB2277845A (en) 1993-05-03 1994-11-09 Philips Electronics Nv Monitoring system
EP0624858A1 (en) 1993-05-14 1994-11-17 Michael Josef Lantschner Method and device for detecting irregular movements of a person
US5519669A (en) 1993-08-19 1996-05-21 At&T Corp. Acoustically monitored site surveillance and security system for ATM machines and other facilities
US5555512A (en) 1993-08-19 1996-09-10 Matsushita Electric Industrial Co., Ltd. Picture processing apparatus for processing infrared pictures obtained with an infrared ray sensor and applied apparatus utilizing the picture processing apparatus
US5712830A (en) 1993-08-19 1998-01-27 Lucent Technologies Inc. Acoustically monitored shopper traffic surveillance and security system for shopping malls and retail space
US5396284A (en) 1993-08-20 1995-03-07 Burle Technologies, Inc. Motion detection system
US5387768A (en) 1993-09-27 1995-02-07 Otis Elevator Company Elevator passenger detector and door control system which masks portions of a hall image to determine motion and court passengers
US5396252A (en) 1993-09-30 1995-03-07 United Technologies Corporation Multiple target discrimination
US5416711A (en) 1993-10-18 1995-05-16 Grumman Aerospace Corporation Infra-red sensor system for intelligent vehicle highway systems
US5473364A (en) 1994-06-03 1995-12-05 David Sarnoff Research Center, Inc. Video technique for indicating moving objects from a movable platform
US5546072A (en) 1994-07-22 1996-08-13 Irw Inc. Alert locator
US6050369A (en) 1994-10-07 2000-04-18 Toc Holding Company Of New York, Inc. Elevator shaftway intrusion device using optical imaging processing
US5666157A (en) 1995-01-03 1997-09-09 Arc Incorporated Abnormality detection and surveillance system
IL116647A (en) 1995-01-03 1999-03-12 Aviv David G Abnormality detection and surveillance system
US6028626A (en) 1995-01-03 2000-02-22 Arc Incorporated Abnormality detection and surveillance system
USRE42690E1 (en) 1995-01-03 2011-09-13 Prophet Productions, Llc Abnormality detection and surveillance system
US5747719A (en) 1997-01-21 1998-05-05 Bottesch; H. Werner Armed terrorist immobilization (ATI) system

Non-Patent Citations (133)

* Cited by examiner, † Cited by third party
Title
"Intelligent Scene Monitoring Drives Security, Surveillance." Signal, Jul. 1995, pp. 29-32.
Agarwal et al. "Estimating Optical Flow from Clustered Trajectory Velocity Time" Pattern Recognition, 1992. vol. I. Conference A: Computer Vision and Applications, Proceedings., 11th IAPR International Conference on Aug. 30-Sep. 3, 1992, pp. 215-219.
Aggarwal et al, "Human Motion Analysis: A Review," Computer Vision and Image Understanding, Mar. 1999, vol. 73, No. 3 pp. 428-440.
Aggarwal et al, "Human Motion Analysis: A Review," the Proceedings IEEE Nonrigid and Articulated Motion Workshop, pp. 90-102 (1997).
Allmen "Image Sequence Description Using Spatiotemporal Flow Curves: Toward Motion-Based Recognition," Thesis submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy Computer Sceinces, University of Wisconsin, Madison, 1991, 153 pages.
Allmen et al., "Computing Spatiotemporal Relations for Dynamic Perceptual Organization." Computer Sciences Department Technical Report 1130, University of Wisconsin-Madison, Dec. 1992, 33 pages.
Arai, Proceedings of the 11th Annual Conference of the Robotics, Society of Japan, No. 1, pp. 347-348 (Nov. 1993).
Ardayfio, David D., Fundamentals of Robotics, Marcel Dekker Inc., (1987).
Aviv "New on-board data processing approach to achieve large compaction," SPIE, 1979, vol. 180, pp. 48-55.
Aviv, D.G. "On Achieving Safer Streets," Library of Congress, TXU 545 919, Nov. 23, 1983, 7 pages.
Aviv, D.G. "The 'Public Eye' Security System," Library of Congress, TXU 551 435, Jan. 11, 1993, 13 pages.
Bergstein et al. "Four-Component Optically Compensated Varifocal System," Journal of the Optical Society of America, Apr. 1962, vol. 52, No. 4, pp. 376-388.
Bergstein et al. "Three-Component Optically Compensated Varifocal System," Journal of the Optical Society of America, Apr. 1962, vol. 52, No. 4, pp. 363-375.
Bergstein et al. "Two-Component Optically Compensated Varifocal System," Journal of the Optical Society of America, Apr. 1962, vol. 52, No. 4, pp. 353-362.
Bilbao et al., "Analysis Procedure of Perimeter Protection Systems: The TDCI Vector," International Carnahan Conference on Security Technology, 1989, pp. 229-235.
Black et al., "Estimating Multiple Independent Motions in Segmented Images using Parametric Models with Local Deformations," Proceedings of the 1994 IEEE Workshop on Motion of Non-Rigid and Articulated Objects (1994). 8 pages.
Bobick, et al., "A State-based Technique for the Summarization and Recognition of Gesture," Fifth Int'l Conf. on Computer Vision, Cambridge, MA, pp. 382-388 (Jun. 1995).
Bolle et al., "Method for Security Monitoring of Undesirable Behavior," IBM Technical Disclosure Bulletin, vol. 37, No. 11 (Nov. 1994). 2 pages.
Bouthemy et al., "Motion Segmentation and Qualitative Dynamic Scene Analysis from an Image Sequence," International Journal of Computer Vision, vol. 10, No. 2, pp. 157-182 (1993).
Brock-Gunn et al., "Using Colour Templates for Target Identification and Tracking," British Machine Vision Conference (1992). 10 pages.
Buker, A neural vision system for robotics applications, Automation, 45, (Oct. 1997). 8 pages.
Buxton et al. "Behavioral descriptions from image sequences," In Proceedings of Workshop on Integration of Natural and Vision Processing Language, 1994, 7 pages.
Buxton et al., "Visual Surveillance in a Dynamic and Uncertain World," Artificial Intelligence vol. 78, pp. 431-359 (1995).
Byrne et al., "Using Shape and Intensity to Track Non-Rigid Objects," University of Leeds School of Computer Studies Research Report Series, Report 94.14, May 1994, 11 pages.
Campbell, et al., "Recognition of Human Body Motion Using Phase Space Constraints," Fifth Int'l. Conf. on Computer Vision, Cambridge, MA, pp. 624-630 (Jun. 1995).
Campbell, et al., "Using Phase Space Constraints to Represent Human Body Motion," Int'l. Workshop on Automatic Face- and Gesture-Recognition, Zurich, pp. 338-343 (1995).
Carvalho et al., "Real Time Automatic Inspection Under Adverse Conditions," SPIE Optical Systems in Adverse Environments, vol. 1399, 1990, pp. 130-136.
Cedras et al. "Motion-based recognition: a survey," Image and Vision Computing, Mar. 1995, vol. 13, No. 2, pp. 129-155.
Chleq et al. "Realtime Image Sequence Interpretation For Video-Surveillance Applications," International Conference on Image Processing, 1996 Proceedings Sep. 16-19, 1996, pp. 801-804.
Custance et al. "Evaluating Scene Monitoring Systems: A Discussion Paper," Security Technology, 199. Crime Countermeasures, Proceedings. Institute of Electrical and Electronics Engineers 1992 International Carnahan Conference on Oct. 14-16, 1992, 7 pages.
Custance et al., "Image Surveillance Systems: Some Novel Design Features," IEEE International Carnahan Conference of Security Technology (1990). 5 pages.
Dance et al., "Interpretation of Dynamic Interaction in Image Sequences," Proceedings of the Artificial Intelligence in Defence Workshop (1995). 15 pages.
Davies et al., "Crowd Monitoring Using Image Processing," IEE Electronic and Communications Engineering Journal, vol. 7, No. 1 pp. 37-47 (Feb. 1995).
Defendants ADT Security Services, Inc., Bosch Security Systems, Inc., Mango DSP, Inc., Pelco, Inc. and Siemens Industry, Inc.'s Invalidity Contentions for U.S. Patent No. RE42,690, Jul. 18, 2012, 59 pages, United States District Court for the Eastern District of Texas Case No. 6: 11-cv-00494-LED.
Devereux et al., "A Method for Evaluating Video Motion Detection," Institute of Electrical and Electronics Engineers 29th Annual International Carnahan Conference on Security Technology (1995). 3 pages.
Electrical Review, Vol. 79 No. 10 pp. 72-76 (Oct. 1994).
Ellis et al. "Model-Based Vision for Automatic Alarm Interpretation," IEEE AES Systems Magazine, Mar. 1991, pp. 14-20.
Ellis et al. "Model-Based Vision for Automatic Alarm Interpretation," Security Technology, 1990. Crime Countermeasures, Proceedings, IEEE 1990 International Carnahan Conference on Oct. 10-12, 1990, pp. 62-67.
Ellis et al., "A Knowledge-Based Approach to Automatic Alarm Interpretation Using Computer Vision on Image Sequences," Proceedings of International Carnahan Conference on Security Technology (1989). 8 pages.
Flinchbaugh, "Robust Video Motion Detection and Event Recognition," Proceedings of DARPA Image Understanding Workshop (May 1997). 4 pages.
Freer et al. "Moving Object Surveillance and Analysis for Camera Based Security Systems," Security Technology, 1995. Proceedings. Institute of Electrical and Electronics Engineers 29th Annual 1995 International Carnahan Conference on Oct. 18-20, 1995, pp. 67-71.
Freer et al., "Automatic Recognition of Suspicious Activity for Camera Based Security Systems," European Convention on Security and Detection, May 1995, Conference Pub. No. 408, pp. 54-58.
Gavrila et al. "Towards 3-D model-based tracking and recognition of human movement: a multi-view approach," International Workshop on Face and Gestrure Recognition, Zurich, 1995, 6 pages.
Gibbins et al. "Detecting Suspicious Background Changes in Video Surveillance of Busy Scenes," Workshop on Applications of Computer Vision, 1996. Dec. 2-4, 1996, 5 pages.
Gould et al., "Detection and Representation of Events in Motion Trajectories," Advances in Image Processing and Analysis, Chapter 14, pp. 393-426 (1992).
Hennebert et al., "Detection of Small and Slow Moving Objects Observed by a Mobile Camera," Theory and Applications of Image Analysis II (1995). pp. 253-266.
Horner "Amethyst: An Enhanced Detection System Intelligently Combining Video Detection and Non-Video Detection Systems," Security Technology, 1995. Proceedings. Institute of Electrical and Electronics Engineers 29th Annual 1995 International Carnahan Conference on Oct. 18-20, 1995, pp. 59-66.
Hosie et al., "Towards Detecting Patterns of Human Behaviour from Image Sequences," Proceedings of the Artificial Intelligence in Defence Workshop (1995). pp. 109-122.
Hötter et al., "Detection of Moving Objects in Natural Scenes by a Stochastic Multi-Feature Analysis of Video Sequences," In Proceedings of Institute of Electrical and Electronics Engineers 29th Annual 1995 International Carnahan Conference on Security Technology (Oct. 1995). pp. 47-52.
Howarth et al. "Selective attention in dynamic vision," Proceedings of the Thirteenth IJCAI Conference, 1993, 7 pages.
Howarth et al., "Analogical Representation of Space and Time," Image and Vision Computing, vol. 10., No. 7 (Sep. 1992). pp. 467-478.
Howell et al., "Video Time Radiation Analysis Program (VTRAP)-Requirements and Preliminary Design Document," (Sep. 1994). 14 pages.
Huttenlocher et al., "Tracking Non-Rigid Objects in Complex Scenes," Proceedings of Fourth International Conference on Computer Vision (May 1993). pp. 93-101.
International Search Report for International (PCT) Patent Application No. PCT/US1996/08674, dated Sep. 17, 1996.
Intille, et al., "Closed-World Tracking," Fifth Int'l. Conf. on Computer Vision, Cambridge, MA, pp. 672-678 (Jun. 1995).
Irani et al., "Detecting and Tracking Multiple Moving Objects Using Temporal Integration," ECCV (1992). 6 pages.
Irani et al., "Motion Analysis for Image Enhancement: Resolution, Occlusion, and Transparency," Journal of Visual Communication and Image Representation, vol. 4, No. 4 pp. 324-335 (1993).
Ju et al., "Cardboard People: A Parameterized Model of Articulated Image Motions," Proceedings of the Second International Conference on Automatic Face and Gesture Recognition (1996). 7 pages.
Kaneta et al., "Image Processing Method for Intruder Detection Around Power Line Towers," IAPR Workshop on Machine Vision Applications, Dec. 1992. pp. 353-356.
Kawashima et al., "Qualitative Image Analysis of Group Behaviour," Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Jun. 1994). pp. 690-693.
Klima et al. "Simple Motion Detection Methods in TV Image for Security Purposes," Security Technology, Proceedings. Institute of Electrical and Electronics Engineers 1993 International Carnahan Conference on Oct. 13-15, 1993, pp. 41-43.
Klima et al., "Motion Detection and Target Tracking in TV Image for Security Purposes," Proceedings of Institute of Electrical and Electronics Engineers 28th Annual International Carnahan Conference on Security Technology (1994). pp. 43-44.
Koga, "A Video Surveillance System Using 3-D Features on Real-Space Coordinates," Proceedings of the 1993 IEICE (The Institute of Electronics, Information and Communication Engineers) Fall Conference, Autumn Pt 6 p. 6.333 (Sep. 1993).
Kollnig et al., "3D Pose Estimation by Fitting Image Gradients Directly to Polyhedral Models," Proceedings of Fifth International Conference on Computer Vision (1995). pp. 569-574.
Kozlow, "David: Advanced Developments for the Next Generation of Video Intrusion Detection," ICCST (1989). pp. 145-147.
Leon et al. "Data fusion for the production of high quality pictures in the automatic sichtprufung," Automatisienrungstechnik, Oct. 1997, pp. 480-489.
Leung et al., "First Sight: A Human Body Outline Labeling System," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 17, No. 4, Apr. 1995, pp. 359-377.
Mackintosh "Sentinel A Technology For Practical Automatic Monitoring Of Busy And Complex Scenes," Security Technology, 1992 Crime Countermeasures, Proceedings. Institute of Electrical and Electronics Engineers 1992 International Carnahan Conference on Oct. 14-16, 1992, pp. 190-193.
MacLean et al., "Recovery of Egomotion and Segmentation of Independent Object Motion Using the EM Algorithm," British Machine Vision Conference, vol. 1 (1994). 11 pages.
Makarov et al., "Intrusion Detection Using Extraction of Moving Edges," Proceedings of the 12th IAPR International Conference on Pattern Recognition (1994). pp. 804-807.
Matas et al., "Constraining Visual Expectations Using a Grammar of Scene Events," Proceedings of the Sixth International Conference on Artificial Intelligence and Information-Control System of Robots (1995). 12 pages.
Matter, "Video Motion Detection for Physical Security Applications," (1990). 14 pages.
McLauchlan et al., "Course Image Motion for Saccade Control," British Machine Vision Conference (1992). 12 pages.
Mecocci et al., "Image Sequence Analysis for Counting in Real Time People Getting In and Out of a Bus," Signal Processing, vol. 35, pp. 105-116 (1994).
Munno et al., "Automatic Video Image Moving Target Detection for Wide Area Surveillance," Institute of Electrical and Electronics Engineers 1993 International Carnahan Conference on Security Technology (1993). pp. 47-57.
Murino et al., "Visual Surveillance by Depth from Focus," 20th International Conference on Industrial Electronics (Sep. 1994). pp. 998-1003.
Murray et al., "Active Exploration of Dynamic and Static Scenes," In Real-Time Computer Vision (1994). 20 pages.
Notice of Allowance for U.S. Appl. No. 08/367,712, mailed Dec. 24, 1996.
Notice of Allowance for U.S. Appl. No. 08/898,470, mailed Mar. 1, 1999.
Notice of Allowance for U.S. Appl. No. 12/466,340, mailed Jul. 18, 2011.
Notice of Allowance for U.S. Appl. No. 12/466,350 mailed Nov. 30, 2011.
Odobez et al., "Detection of Multiple Moving Objects Using Multiscale MRF with Camera Motion Compensation," 1st ICIP (Nov. 1994). 5 pages.
Official Action for U.S. Appl. No. 08/367,712, mailed Jul. 24, 1996.
Official Action for U.S. Appl. No. 08/898,470, mailed Oct. 1, 1998.
Official Action for U.S. Appl. No. 12/466,340, mailed Apr. 18, 2011.
Official Action for U.S. Appl. No. 12/466,340, mailed Aug. 30, 2010.
Official Action for U.S. Appl. No. 12/466,340, mailed Mar. 11, 2010.
Official Action for U.S. Appl. No. 12/466,340, mailed Nov. 8, 2010.
Official Action for U.S. Appl. No. 12/466,350 mailed Dec. 22, 2010.
Official Action for U.S. Appl. No. 12/466,350 mailed Mar. 15, 2010.
Official Action for U.S. Appl. No. 13/361,438, mailed Sep. 5, 2012 5 pages.
Pearce et al., "Rulegraphs for Graph Matching in Pattern Recognition," Pattern Recognition, vol. 27, No. 9, pp. 1231-1247 (1994).
Pentland et al., "Recovery of Nonrigid Motion and Structure," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 13, No. 7 (Jul. 1991). pp. 730-742.
Polana et al., "Low Level Recognition of Human Motion (Or How to get Your Man Without Finding his Body Parts)," Proceedings of IEEE Computer Society Workshop on Motion of Non-Rigid and Articulated Objects (1994). 6 pages.
Proceedings of JSPE (The Japan Society for Precision Engineering) Spring Conference, No. 3 p. 891-892 (Jul. 1993).
Rabiner "Applications of Voice Processing to Telecommunications," Proceedings of the IEEE, Feb. 1994, vol. 82, No. 2, pp. 199-228.
Rabiner "The Role of Voice Processing in Telecommunications," 2nd IEEE Workshop on Interactive Voice Technology for Telecommunications Applications (IVTTA94) Sep. 1994, 8 pages.
Rabiner et al. "Fundamental of Speech Recognition," Prentice Hall International, Inc., Apr. 12, 1993, pp. 434-495.
Rabiner et al., "A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition," Readings in Speech Recognition (1990). pp. 267-296.
Rabiner et al., Fundamentals of Speech Recognition, Chapters 8-9 (1993). 64 pages.
Rangarajan et al., "Matching Motion Trajectories Using Scale-Space," Pattern Recognition, vol. 26, No. 4, pp. 595-610 (1993).
Retz-Schmidt, "A Replai of Soccer: Recognizing Intentions in the Domain of Soccer Games," European Conference on Artificial Intelligence (Aug. 1988). pp. 455-457.
Richardson, "A Sequential Detection Approach to Target Tracking," A thesis submitted to the Department of Electrical Engineering for the degree of Master of Science, Queen's University, Kingston, Ontario, Canada (Jun. 1992). 177 pages.
Robinson "Neural Network Solutions Provide Facial Recognition," Signal, Feb. 1991, vol. 45, No. 6, pp. 73-76.
Rodger et al., "Video Motion Detection Systems: A Review for the Nineties," ("Rodger"), Proceedings of Institute of Electrical and Electronics Engineers 28th Annual International Carnahan Conference on Security Technology (1994). pp. 92-97.
Rohr "Towards Model-Based Recognition of Human Movements in Image Sequences," CVGIP: Image Understanding, Jan. 1994, vol. 59, No. 1, pp. 94-115.
Sakuma, "Detecting intruders using the method of Inter Frame Difference," 1990 Autumn National Convention Record, The Institute of Electronics, Information and Communication Engineers, pt. 6 (Sep. 1990).
Sakuma, "Detecting intruders using the method of Inter Frame Difference," ITEJ (The Institute of Television Engineers of Japan) Technical Report, vol. 14 No. 49 (IPCV90 27-30/AIPS90 50-53), pp. 1-6 (Sep. 1990).
Seki et al., "A Highly Reliable Intruder Monitoring System that uses Movement Information," Mitsubishi Denki Giho, vol. 67, No. 7 (1993). 16 pages with English Translation.
Shimonaga, "Automatic Supervisory of Intrusion with ITV," Proceedings of the 1992 ITE (The Institute of Television Engineers) Annual Convention, pp. 351-352 (Jul. 1992). (English abstract).
Shio et al. "Segmentation of People in Motion," Visual Motion, 1991., Proceedings of the IEEE Workshop on Oct. 7-9, 1991, pp. 325-332.
Smith, "Asset-2: Real-Time Motion Segmentation and Shape Tracking," IEEE Fifth International Conference on Computer Vision (Jun. 20, 1995). pp. 237-244.
Starner et al., "Real-Time American Sign Language Recognition from Video Using Hidden Markov Models," M.I.T. Media Laboratory Perceptual Computing Section Technical Report No. 375(1995). 7 pages.
Stiller, Computer-Age, Digital-Color Camera COMPRO von Pearl Agency (1996). 5 pages German only.
Stubbington, "Intelligent Scene Monitoring; Technical Aspects and Practical Experience," Security Technology, Carnahan Conference, Oct. 18-20, 1995, pp. 364-375.
Suzuki et al. "Extracting Non-Rigid Moving Objects by Temporal Edges,"Pattern Recognition, 1992. vol. I. Conference A: Computer Vision and Applications, Proceedings., 11th IAPR International Conference on Aug. 30-Sep. 3, 1992, pp. 69-73.
Takano et al., "Intruder Detection System by Image Processing," Institute of Electrical and Electronics Engineers 28th Annual 1994 International Carnahan Conference on Security Technology (1994). pp. 31-33.
Takatoo et al. "Detection of Objects Including Persons Using Image Processing," Pattern Recognition, 1996, Proceedings of the 13th International Conference on Aug. 25-29, 1996, pp. 466-472.
Technical Research Report of Shimizu Corporation, vol. 60, pp. 123-131 (Oct. 1994). Japanese only.
Toal et al., "Spatio-temporal Reasoning within a Traffic Surveillance System," Proceeds of the Second European Conference on Computer Vision (1992). pp. 884-892.
Tsuge et al., "Accident Vehicle Automatic Detection System by Image Processing Technology," Vehicle Navigation & Information Systems Conference Proceedings (1994). pp. 45-50.
U.S. Appl. No. 12/466,350, filed May 14, 2009, Aviv.
U.S. Appl. No. 13/361,438, filed Jan. 30, 2012, Aviv.
Weibel et al. "Readings in Speech Recognition," Morgan Kaufam, May 15, 1990 pp. 267-296.
Wilson, et al., "Configuration States for the Representation and Recognition of Gesture," Int'l. Workshop on Automatic Face- and Gesture-Recognition, Zurich, pp. 129-134 (1995).
Wollert, Get the picture, Image Processing Industrial Market Report, Elektronik (Feb. 1997). 12 pages German Only.
Wren et al., "Pfinder: Real-Time Tracking of the Human Body," 1996 Proceedings of the 2nd International Conference on Automatic Face and Gesture Recognition, (1996). pp. 51-56.
Wren et al., "Pfinder: Real-Time Tracking of the Human Body," IEEE Transactions on Pattern Analysis and Machine Intelligence (1997). pp. 780-785.
Xiao et al. "Eleview: An Active Elevator Monitoring Vision System," MVA 96, IAPR Workshop on Machine Vision Applications, in Tokyo, Japan, Nov. 12, 1996, pp. 253-256.
Yamato et al., "Recognizing Human Action in Time-Sequential Images using Hidden Markov Model," Proceedings of IEEE Computer Society Conference on Computer Vision (Jun. 1992). pp. 379-385.
Yamato et al., "Recognizing Human Action in Time-Sequential Images using Hidden Markov Model," The Transactions of The Institute of Electronics, Information and Communication Engineers, vol. 76, No. 12, Dec. 1993, pp. 2556-2563, Japanese Only.
Yeh et al. "A Vision System for Safe Robot Operation," 1988 IEEE International Conference on Robotics and Automation, Apr. 24-29, 1988, pp. 1461-1465.
Yoshikawa et al., "Development of Video Surveillance System using Motion Information," Mitsubishi Electric Corp. (1991). 5 pages Japanese Only.

Cited By (89)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10949773B2 (en) 2005-10-26 2021-03-16 Cortica, Ltd. System and methods thereof for recommending tags for multimedia content elements based on context
US11019161B2 (en) 2005-10-26 2021-05-25 Cortica, Ltd. System and method for profiling users interest based on multimedia content analysis
US11758004B2 (en) 2005-10-26 2023-09-12 Cortica Ltd. System and method for providing recommendations based on user profiles
US20170262438A1 (en) * 2005-10-26 2017-09-14 Cortica, Ltd. System and method for determining analytics based on multimedia content elements
US11061933B2 (en) 2005-10-26 2021-07-13 Cortica Ltd. System and method for contextually enriching a concept database
US11620327B2 (en) 2005-10-26 2023-04-04 Cortica Ltd System and method for determining a contextual insight and generating an interface with recommendations based thereon
US11032017B2 (en) 2005-10-26 2021-06-08 Cortica, Ltd. System and method for identifying the context of multimedia content elements
US11922293B2 (en) 2005-10-26 2024-03-05 Cortica Ltd. Computing device, a system and a method for parallel processing of data streams
US11361014B2 (en) 2005-10-26 2022-06-14 Cortica Ltd. System and method for completing a user profile
US11604847B2 (en) 2005-10-26 2023-03-14 Cortica Ltd. System and method for overlaying content on a multimedia content element based on user interest
US10742340B2 (en) 2005-10-26 2020-08-11 Cortica Ltd. System and method for identifying the context of multimedia content elements displayed in a web-page and providing contextual filters respective thereto
US10848590B2 (en) 2005-10-26 2020-11-24 Cortica Ltd System and method for determining a contextual insight and providing recommendations based thereon
US11386139B2 (en) 2005-10-26 2022-07-12 Cortica Ltd. System and method for generating analytics for entities depicted in multimedia content
US11238066B2 (en) 2005-10-26 2022-02-01 Cortica Ltd. Generating personalized clusters of multimedia content elements based on user interests
US11657079B2 (en) 2005-10-26 2023-05-23 Cortica Ltd. System and method for identifying social trends
US10902049B2 (en) 2005-10-26 2021-01-26 Cortica Ltd System and method for assigning multimedia content elements to users
US11403336B2 (en) 2005-10-26 2022-08-02 Cortica Ltd. System and method for removing contextually identical multimedia content elements
US11216498B2 (en) 2005-10-26 2022-01-04 Cortica, Ltd. System and method for generating signatures to three-dimensional multimedia data elements
US11232326B2 (en) 2007-07-03 2022-01-25 Shoppertrak Rct Corporation System and process for detecting, tracking and counting human objects of interest
US11537636B2 (en) 2007-08-21 2022-12-27 Cortica, Ltd. System and method for using multimedia content as search queries
US10909695B2 (en) * 2010-01-11 2021-02-02 Shoppertrak Rct Corporation System and process for detecting, tracking and counting human objects of interest
US20180300887A1 (en) * 2010-01-11 2018-10-18 Shoppertrak Rct Corporation System and process for detecting, tracking and counting human objects of interest
US20120268269A1 (en) * 2011-04-19 2012-10-25 Qualcomm Incorporated Threat score generation
US20160133122A1 (en) * 2013-06-25 2016-05-12 Nec Corporation Sensitivity adjustment device, sensitivity adjustment method, storage medium, and monitoring system
US9084411B1 (en) 2014-04-10 2015-07-21 Animal Biotech Llc Livestock identification system and method
US20170275134A1 (en) * 2014-11-26 2017-09-28 Otis Elevator Company Elevator security and control system based on passenger movement
US10696517B2 (en) * 2014-11-26 2020-06-30 Otis Elevator Company Elevator security and control system based on passenger movement
US9984154B2 (en) 2015-05-01 2018-05-29 Morpho Detection, Llc Systems and methods for analyzing time series data based on event transitions
US10839009B2 (en) 2015-05-01 2020-11-17 Smiths Detection Inc. Systems and methods for analyzing time series data based on event transitions
US11195043B2 (en) 2015-12-15 2021-12-07 Cortica, Ltd. System and method for determining common patterns in multimedia content elements based on key points
US11037015B2 (en) 2015-12-15 2021-06-15 Cortica Ltd. Identification of key points in multimedia data elements
US11783613B1 (en) 2016-12-27 2023-10-10 Amazon Technologies, Inc. Recognizing and tracking poses using digital imagery captured from multiple fields of view
US10839203B1 (en) 2016-12-27 2020-11-17 Amazon Technologies, Inc. Recognizing and tracking poses using digital imagery captured from multiple fields of view
US10699421B1 (en) 2017-03-29 2020-06-30 Amazon Technologies, Inc. Tracking objects in three-dimensional space using calibrated visual cameras and depth cameras
US11315262B1 (en) 2017-03-29 2022-04-26 Amazon Technologies, Inc. Tracking objects in three-dimensional space using calibrated visual cameras and depth cameras
US11760387B2 (en) 2017-07-05 2023-09-19 AutoBrains Technologies Ltd. Driving policies determination
US11899707B2 (en) 2017-07-09 2024-02-13 Cortica Ltd. Driving policies determination
US11232294B1 (en) 2017-09-27 2022-01-25 Amazon Technologies, Inc. Generating tracklets from digital imagery
US11861927B1 (en) 2017-09-27 2024-01-02 Amazon Technologies, Inc. Generating tracklets from digital imagery
US11284041B1 (en) 2017-12-13 2022-03-22 Amazon Technologies, Inc. Associating items with actors based on digital imagery
US11030442B1 (en) * 2017-12-13 2021-06-08 Amazon Technologies, Inc. Associating events with actors based on digital imagery
US11468681B1 (en) 2018-06-28 2022-10-11 Amazon Technologies, Inc. Associating events with actors using digital imagery and machine learning
US11468698B1 (en) 2018-06-28 2022-10-11 Amazon Technologies, Inc. Associating events with actors using digital imagery and machine learning
US11482045B1 (en) 2018-06-28 2022-10-25 Amazon Technologies, Inc. Associating events with actors using digital imagery and machine learning
US11922728B1 (en) 2018-06-28 2024-03-05 Amazon Technologies, Inc. Associating events with actors using digital imagery and machine learning
US10846544B2 (en) 2018-07-16 2020-11-24 Cartica Ai Ltd. Transportation prediction system and method
US11613261B2 (en) 2018-09-05 2023-03-28 Autobrains Technologies Ltd Generating a database and alerting about improperly driven vehicles
US11282391B2 (en) 2018-10-18 2022-03-22 Cartica Ai Ltd. Object detection at different illumination conditions
US11029685B2 (en) 2018-10-18 2021-06-08 Cartica Ai Ltd. Autonomous risk assessment for fallen cargo
US10839694B2 (en) 2018-10-18 2020-11-17 Cartica Ai Ltd Blind spot alert
US11087628B2 (en) 2018-10-18 2021-08-10 Cartica Al Ltd. Using rear sensor for wrong-way driving warning
US11718322B2 (en) 2018-10-18 2023-08-08 Autobrains Technologies Ltd Risk based assessment
US11417216B2 (en) 2018-10-18 2022-08-16 AutoBrains Technologies Ltd. Predicting a behavior of a road used using one or more coarse contextual information
US11685400B2 (en) 2018-10-18 2023-06-27 Autobrains Technologies Ltd Estimating danger from future falling cargo
US11673583B2 (en) 2018-10-18 2023-06-13 AutoBrains Technologies Ltd. Wrong-way driving warning
US11373413B2 (en) 2018-10-26 2022-06-28 Autobrains Technologies Ltd Concept update and vehicle to vehicle communication
US11270132B2 (en) 2018-10-26 2022-03-08 Cartica Ai Ltd Vehicle to vehicle communication and signatures
US11700356B2 (en) 2018-10-26 2023-07-11 AutoBrains Technologies Ltd. Control transfer of a vehicle
US11904863B2 (en) 2018-10-26 2024-02-20 AutoBrains Technologies Ltd. Passing a curve
US11170233B2 (en) 2018-10-26 2021-11-09 Cartica Ai Ltd. Locating a vehicle based on multimedia content
US11392738B2 (en) 2018-10-26 2022-07-19 Autobrains Technologies Ltd Generating a simulation scenario
US11244176B2 (en) 2018-10-26 2022-02-08 Cartica Ai Ltd Obstacle detection and mapping
US11126869B2 (en) 2018-10-26 2021-09-21 Cartica Ai Ltd. Tracking after objects
US10789535B2 (en) 2018-11-26 2020-09-29 Cartica Ai Ltd Detection of road elements
US10559189B1 (en) 2018-12-14 2020-02-11 John Dante Menna System, method and apparatus for providing voice alert notification of an incident
US11170647B2 (en) 2019-02-07 2021-11-09 Cartica Ai Ltd. Detection of vacant parking spaces
US11643005B2 (en) 2019-02-27 2023-05-09 Autobrains Technologies Ltd Adjusting adjustable headlights of a vehicle
US11285963B2 (en) 2019-03-10 2022-03-29 Cartica Ai Ltd. Driver-based prediction of dangerous events
US11755920B2 (en) 2019-03-13 2023-09-12 Cortica Ltd. Method for object detection using knowledge distillation
US11694088B2 (en) 2019-03-13 2023-07-04 Cortica Ltd. Method for object detection using knowledge distillation
US11132548B2 (en) 2019-03-20 2021-09-28 Cortica Ltd. Determining object information that does not explicitly appear in a media unit signature
US10846570B2 (en) 2019-03-31 2020-11-24 Cortica Ltd. Scale inveriant object detection
US11727056B2 (en) 2019-03-31 2023-08-15 Cortica, Ltd. Object detection based on shallow neural network that processes input images
US11741687B2 (en) 2019-03-31 2023-08-29 Cortica Ltd. Configuring spanning elements of a signature generator
US11481582B2 (en) 2019-03-31 2022-10-25 Cortica Ltd. Dynamic matching a sensed signal to a concept structure
US10748038B1 (en) 2019-03-31 2020-08-18 Cortica Ltd. Efficient calculation of a robust signature of a media unit
US11488290B2 (en) 2019-03-31 2022-11-01 Cortica Ltd. Hybrid representation of a media unit
US11275971B2 (en) 2019-03-31 2022-03-15 Cortica Ltd. Bootstrap unsupervised learning
US11908242B2 (en) 2019-03-31 2024-02-20 Cortica Ltd. Efficient calculation of a robust signature of a media unit
US11704292B2 (en) 2019-09-26 2023-07-18 Cortica Ltd. System and method for enriching a concept database
US11593662B2 (en) 2019-12-12 2023-02-28 Autobrains Technologies Ltd Unsupervised cluster generation
US11590988B2 (en) 2020-03-19 2023-02-28 Autobrains Technologies Ltd Predictive turning assistant
US11825221B2 (en) 2020-03-31 2023-11-21 Western Digital Technologies, Inc. Sensor system with low power sensor devices and high power sensor devices
US11233956B2 (en) 2020-03-31 2022-01-25 Western Digital Technologies, Inc. Sensor system with low power sensor devices and high power sensor devices
US11827215B2 (en) 2020-03-31 2023-11-28 AutoBrains Technologies Ltd. Method for training a driving related object detector
US11954168B2 (en) 2020-03-31 2024-04-09 Cortica Ltd. System and method thereof for dynamically associating a link to an information resource with a multimedia content displayed in a web-page
US11443516B1 (en) 2020-04-06 2022-09-13 Amazon Technologies, Inc. Locally and globally locating actors by digital cameras and machine learning
US11398094B1 (en) 2020-04-06 2022-07-26 Amazon Technologies, Inc. Locally and globally locating actors by digital cameras and machine learning
US11750639B2 (en) 2021-04-05 2023-09-05 Bank Of America Corporation ATM-based anomaly and security threat detection

Also Published As

Publication number Publication date
USRE42690E1 (en) 2011-09-13
WO1999005867A1 (en) 1999-02-04
WO1999005867A9 (en) 1999-05-06
WO1999005867B1 (en) 1999-04-01
US6028626A (en) 2000-02-22
EP0997041A1 (en) 2000-05-03

Similar Documents

Publication Publication Date Title
USRE44225E1 (en) Abnormality detection and surveillance system
US5666157A (en) Abnormality detection and surveillance system
Elharrouss et al. A review of video surveillance systems
US7683929B2 (en) System and method for video content analysis-based detection, surveillance and alarm management
US20040240542A1 (en) Method and apparatus for video frame sequence-based object tracking
Crocco et al. Audio surveillance: A systematic review
EP3682429B1 (en) System and method for gate monitoring during departure or arrival of an autonomous vehicle
KR101644443B1 (en) Warning method and system using prompt situation information data
CN101119482B (en) Overall view monitoring method and apparatus
EP1218867B1 (en) Improvements in and relating to surveillance systems
US9247215B1 (en) Laser sensor system
JP2000244897A (en) State recognition system and state recognition display generation method
US10997430B1 (en) Dangerous driver detection and response system
Gao et al. Pattern of life from WAMI objects tracking based on visual context-aware tracking and infusion network models
SG191430A1 (en) System and method for parallel image processing
CN105100700A (en) Target tracking device using handover between cameras and method thereof
US8860812B2 (en) Ambient presentation of surveillance data
CN107360394A (en) More preset point dynamic and intelligent monitoring methods applied to frontier defense video monitoring system
WO2021095351A1 (en) Monitoring device, monitoring method, and program
KR20190035958A (en) Situation ascertainment system using camera picture data, control device, and situation ascertainment method using camera picture data
Ho et al. Public space behavior modeling with video and sensor analytics
Flammini et al. Challenges and emerging paradigms for augmented surveillance
US20240062636A1 (en) System and method for identifying a weapon, and tracking and generating an alert in response to the weapon identification
KR20160086536A (en) Warning method and system using prompt situation information data
Francisco et al. Thermal imaging for law enforcement and security: post 9-11