WO2006021943A1 - Apparatus and method for multimedia content based - Google Patents

Apparatus and method for multimedia content based Download PDF

Info

Publication number
WO2006021943A1
WO2006021943A1 PCT/IL2004/000736 IL2004000736W WO2006021943A1 WO 2006021943 A1 WO2006021943 A1 WO 2006021943A1 IL 2004000736 W IL2004000736 W IL 2004000736W WO 2006021943 A1 WO2006021943 A1 WO 2006021943A1
Authority
WO
WIPO (PCT)
Prior art keywords
content
masking
interaction
metadata
masked
Prior art date
Application number
PCT/IL2004/000736
Other languages
French (fr)
Inventor
Shay Gabay
Doron Sitbon
Dafna Levy Philosoph
Shai Shermister
Original Assignee
Nice Systems Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=35967195&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=WO2006021943(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Nice Systems Ltd. filed Critical Nice Systems Ltd.
Priority to US10/507,582 priority Critical patent/US7714878B2/en
Priority to PCT/IL2004/000736 priority patent/WO2006021943A1/en
Publication of WO2006021943A1 publication Critical patent/WO2006021943A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/1066Session management
    • H04L65/1101Session protocols
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/70Media network packetisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • H04L65/764Media network packet handling at the destination 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/56Provisioning of proxy services
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/56Provisioning of proxy services
    • H04L67/564Enhancement of application control based on intercepted application data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4318Generation of visual interfaces for content selection or interaction; Content or additional data rendering by altering the content in the rendering process, e.g. blanking, blurring or masking an image region
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/4508Management of client data or end-user data
    • H04N21/4532Management of client data or end-user data involving end-user characteristics, e.g. viewer profile, preferences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments

Definitions

  • the present invention relates to US patent application serial number
  • the present invention relates to content derived systems, in general and
  • interactions could include the quality assurance of agent's performance, security
  • call centers use capture
  • the entire interaction In a first example, if the interaction is a telephone call, the
  • supervisor listening to the interaction can listen to the entire phone call and to all
  • access rights to view the video can view the entire video footage which comprises
  • a supervisor having access rights to view employee's e-mails will be able to view entire set of
  • Interactions may comprise a number of
  • speakers a plurality of persons appearing within a video footage or stream
  • One aspect of the present invention regards an apparatus for content based
  • the apparatus comprises a
  • a second aspect of the present invention regards a method for content
  • the method comprises generating content
  • control proposed by the present invention is associated with a portion of
  • a third aspect of the present invention includes a method for content
  • said stream comprising one or
  • the method comprising generating content masking on at least a
  • the method further comprises the step ?
  • a content input device can be a camera or a microphone or other data capturing
  • the method can further comprise the step of masking the metadata
  • a display device can be any one of the following: a display device, a computer screen, a
  • content masking in association with the interaction can be altered based on one or
  • the method can further comprise the step of analyzing the
  • the method further comprises the step of selecting one or more content
  • method further comprises the step of altering at least a part of the interaction to be
  • the step of generating content masking can comprise
  • content masking can also comprise generating content masking from a first
  • each said masked content is
  • the step of generating content masking can also be
  • the content can be any multimedia content.
  • the user data can be any
  • Masking can include any one of the following:
  • masking generator for generating content masking on at least a part of the
  • the apparatus further comprises a metadata capture device for
  • the apparatus can also capture video, audio or screen stream.
  • the apparatus can also capture video, audio or screen stream.
  • the apparatus can also capture video, audio or screen stream.
  • the apparatus further comprises
  • an event identification device to recognize events in content associated with the
  • streaming device to generate selectively masked content associated with the
  • the masking content generators alter at least a part of the interaction
  • the apparatus can further comprise a masked storage to store masked content and an unmasked storage to store
  • the apparatus further comprises a triggering device to activate
  • Fig. 1 is a schematic block diagram of the multimedia content based
  • Fig. 2 is a schematic block diagram of the components of the
  • Fig. 3 is a flow chart describing the steps of a method for multimedia
  • Fig. 4 is a flow chart describing the steps of a method for multimedia
  • Pan-Zoom-Tilt spectrum change, frame rate change, mute sound, and the like.
  • An interaction can comprise one of
  • the proposed apparatus and method is designed to enable or restrict
  • proposed apparatus and method is further designed to mask, according to its
  • the masking can be done either in a negative or a
  • masking can be used to mask the entire content except a
  • the masking could be applied to the information stream itself or on the
  • enhancement is based on the content of the interaction, on other metadata
  • the masking is performed based on various parameters
  • predetermined events and provide predetermined actions for masking (allowing or
  • screen streams comprising one or more interactions are captured and content masking is generated on at least a part of the interaction.
  • the content masking is
  • Metadata associated with the interaction is obtained from a content input device
  • Alteration means the changing of the masking applied to the
  • the metadata is associated with the segment of the interaction for which content
  • apparatus for multimedia content based masking generally referenced 100.
  • apparatus 100 comprises a multimedia capture device 102, a metadata capture
  • the apparatus 100 could also comprise a manual
  • masking generator 116 a masking methods engine 118, a tuning engine 120, a
  • apparatus 100 can also comprise one or more data structures to handle the
  • an events definition data structure 126 includes an events definition data structure 126, an events interaction data structure 128, a user administration data structure 130, and a masking quality data structure
  • various storage devices can be associated with the apparatus 100.
  • Such storage devices can further comprise a masked /unmasked multimedia
  • the apparatus 100 is connected or can comprise
  • the multimedia capture devices 102 include one or more multimedia capture devices 102.
  • capture devices can include capture devices capturing voice, video, screen events, screens, key
  • a voice capturing device can be a
  • radio frequency receiving device a voice over IP module or device, a facsimile
  • a video capture can include the stream captured or the
  • the capture device can also include a computing device, mobile or
  • Such can include any data
  • - Physical access devices can include "smart" access restricting devices or barriers
  • doors which can be opened or locked or controlled from a remote location, cameras or microphone or other recording or logging equipment
  • any other such similar physical device which can be controlled from a distance
  • devices can include devices for controlling of public transportation vehicles, such as
  • Such vehicles like crossings, traffic lights, semaphores, landing lights or landing
  • controlled devices can include heating and cooling systems as well as air
  • sensors could be captured and processed.
  • data from a chemical sensor could be captured and processed.
  • optical sensor could provide motion indications and page detections.
  • Mechanical sensors could be associated with scales or various mechanical
  • the present invention envisage the capturing of such an interaction by one or more
  • capture devices capturing one or more aspects, such as voice, video, data and the
  • the capture devices can be a directly
  • Locations can include buildings and any combination
  • the apparatus 100 also comprises a metadata capture
  • the metadata received is associated with the multimedia captured and
  • the multimedia capture device 102 can include the location of the multimedia capture device 102, or the location of
  • the metadata can include data relating to the multimedia from other systems.
  • data can include information stating from which recording systems or logging
  • Metadata can include environmental conditions
  • example can be information about a person or a location shown within a video
  • Another non-limiting example could include data indicating the
  • example could include GPS (Global Positioning System) data indicating'
  • Metadata capture device 104 can be a constituent part of the multimedia capture
  • a camera capturing a scene can also transmit the
  • the metadata capture device 104 may also receive data or information from
  • CTI Computer Telephony Interface
  • CDR Detail Recording
  • IVR Interactive Voice Response
  • DVR Dual Tune Multi
  • analysis 106 uses various analysis on the data received through the use of content
  • the content analysis engines 122 is linked on the one side to
  • multimedia capture device 102 and the metadata capture device 104 and
  • engines 122 could be allocated statically or dynamically in order to optimize
  • the content analysis engines 122 could be available for online
  • 122 is based on metadata capture of offline information, such as, for example,
  • content analysis engines 122 could include a word spotting content analysis
  • a prefix For example, absence of words can be used to later identify a segment of the application missing which means that a specific interaction cannot be released
  • the content analysis engines 122 can also include a speech
  • the content analysis engines 122 can also include a video analysis ' "
  • the video analysis engine can also determine the velocity of objects, such as !
  • the content analysis engines 122 could also be used to analyze the content of the content.
  • the content analysis engines 122 could also be used to analyze the content of the content.
  • the content analysis engines 122 can also include
  • a screen analysis engine to analyze activities performed on the screen of a
  • IVR analysis to identify the various IVR elements.
  • the various IVR elements IVR analysis to identify the various IVR elements.
  • content analysis engines 122 can comprise an audio analysis engine to determine
  • emotion detection to detect high emotions of one of the speakers, speaker hunting
  • gender detection to identify the gender of the speakers and silence detection to
  • emotion detector can detect the level of the volume of the speaker's voice and
  • classify the emotion such as whether the person is nervous or upset and the like.
  • the content analysis engines 122 can also comprise a key stroke
  • the content analysis engines 122 can comprise data analysis '
  • the apparatus of the present invention may restrict access to a video interaction.
  • One non-limiting example relates to a situation in
  • FOV of view
  • the face of a particular person, such as a client may be masked to avoid his or her
  • the face region of interest can be enhanced while other elements are
  • Another example relates to words uttered or written
  • employee's salary information will be deleted, hidden, concealed or masked.
  • An additional non-limiting example concerns a marketing manager or a product manager that are enabled to access only those segments of an
  • masking operation can be associated with both the hiding or blurring of
  • An event is a number of data types, such as events, actions and layers.
  • An event is a number of data types, such as events, actions and layers.
  • One or more events can comprise a scenario of various activities.
  • a scenario can be an interaction in which credit card or personal details "
  • the credit card scenario could involve a set of consecutive events
  • Such events could include credit card detail
  • example is an interaction in which an agent advises a customer as to possible
  • a scenario may comprise one or more events
  • plurality of events can characterize a scenario. Events could be identified during
  • multimedia content retrieval request is made in order perform content playback
  • Non-limiting examples can include be a
  • a defined event can comprise one or more -
  • one event can be the credit card details screen being loaded by the
  • key strokes including a template of a credit card, such as
  • events can be defined as a
  • Compound events may be based on the existence or non-existence of "progeny" or
  • the events definition data structure 126 comprises actions. Actions are
  • the data structure 126 comprises a
  • Layers are a series of actions associated with an event.
  • the action in layer 1 can be "mute the
  • the data structure 126 also comprises data logical links
  • the layer can be associated with the profile of a user. For example, a layer titled
  • “medical details” can be defined as the actions replacing characters on a screen.
  • the events identifications engine 108 processes the metadata
  • the events identifications engine 108 can use a number of alternative
  • module 108 could perform a single pass over the metadata in order to identify one
  • module 108 can perform multiple passes over the same metadata in order to
  • compound events are events for which a
  • passes may be on the same metadata or on metadata associated with different
  • An example of a compound event is an event comprising the existence
  • engine 108 stores the spatial grid location of the event identified. Such can be a
  • events identification engine 108 so as to identify the item or object or event
  • an event comprises "locate persons X"
  • recognition content analysis engine includes the identity of person X on a number
  • the engine 108 will retrieve said frames and use the event associated
  • the engine 108 will identify the spatial coordinates on each
  • interaction data structure 128 for later use.
  • the action can call for
  • the engine 108 can search events or scenarios on
  • non-limiting example can be a search to be performed every 30 seconds or another
  • the engine 108 generates a data record for each metadata for
  • the data record includes the identification of the
  • data fields may include the associated layer of the actions, the scenarios names,
  • the data records associated with the identified events are stored in the events
  • the engine 108 can optionally mark the interaction in which an event has been identified. To mark the interaction the
  • the engine 108 stores a status flag on the interaction stream.
  • the status flag is
  • said data fields comprising the data record associated with the data
  • the mark could be performed by adding a data bit or a number of data
  • the data bit comprises a Boolean value.
  • the data bit can alternatively
  • the playback device to quickly determine if the person playing back the interaction
  • Such data structure includes a table comprising the event handle
  • the request/push handler 110 will also handle requests
  • handler 110 is responsible for
  • a supervisor may issue a request through the handler 110 to playback the
  • request/push handler 110 which retrieves the said request interaction, the events
  • the request/push handler 110 can generate metadata that could be used for
  • a video or and audio segment that is
  • the rule will affect the marking of the segment for a manual review by an
  • User administrative data structure 130 holds
  • the retrieval request can be for any type of interaction
  • masking generator 112 in association with the masking methods engine 118 is
  • the masking method engine 118 effecting the provision of a suitable multimedia
  • the multimedia content masking method could be any multimedia content masking method.
  • the multimedia content masking method could be any multimedia content masking method.
  • the video masking method could include video block
  • masking sub-method could include the following block masking options: a) denial to replay a request, b) implementation of a black screen during play, c)
  • masked stream elements such as gain level, storage options, and the like.
  • video concealing sub-method could include the following concealment options: a)
  • the voice masking method includes a silence implementation sub-method, a
  • the search masking sub-method could include mask search of the record,
  • ERP Enterprise Resource
  • CRM Customer Relationship Management
  • PLM personal Liaison
  • the screen masking method includes screen block
  • block masking sub-method includes a) denial of the play request, b)
  • the screen concealing sub-method includes the insertion
  • the masking generator 112 relays information concerning the type
  • the masking quality data structure 132 stores masking
  • Data structure 132 could be tuned optionally by the tuning
  • the masked multimedia is streamed by the masked multimedia streaming 114 and the masked
  • multimedia is presented to the user as output 124.
  • multimedia could be stored in the masked multimedia storage 134 for a pre ⁇
  • containing financial negotiation could be replaced by a replacement screen, such
  • a masking method could be enhanced via the utilization of a
  • the generation of the masking formula could be either
  • ROI region-of-interest
  • the ROI is defined on the video stream by the utilization of Field of View (FOV) limitation.
  • FOV Field of View
  • one of the solutions is to reduce the size of the ROI.
  • the reduction could be made semi-automatically by recommending to the user an
  • the mask formula will include an improved masking algorithm that
  • Multimedia elements to be masked in accordance with the
  • masking formula could include entire management interactions, (phone calls with
  • handlers could be co-located on the same computing platform.
  • a multiplexing device could be added in
  • a de-multiplexer unit could be placed between the
  • FIG. 2 an apparatus 200 is shown in accordance with
  • the apparatus 200 includes an interaction handler 202, an events definition handler and identifier 204, a
  • the interaction handler 202 manages the multimedia interactions
  • interaction could be performed directly face-to-face or could be performed via
  • interaction involves the capturing of the interaction data via diverse multimedia
  • the interaction management further includes the capturing of the
  • Metadata includes control information embedded in the interaction data.
  • Metadata could include indications concerning DTMF signaling, e-mail addresses,
  • Metadata could be captured separately or could be extracted from the interaction
  • the captured metadata could include locations, time and date, participants' identification, priority parameters, CDR/CTI information IVR
  • content is captured in a multimedia format, such as video, audio, graphics, text,
  • multimedia content is analyzed in order to identify the control information carried
  • the metadata analysis could include diverse analysis techniques, such as
  • word spotting speech-to-text, emotion detection, video analysis, screen analysis,
  • the multimedia content is
  • identifier 204 processes the multimedia stream in order to identify, to recognize
  • the requests/push handler 206 is responsible for both the automatic initiation of
  • the generator 212 collects
  • the masking method selector module 208 is a masking method selector module 208.
  • the masking generator 212 in association with the masking quality data handler 210,
  • the multimedia stream in order to conceal specifically pre-defined information
  • the masking could be performed manually by the user
  • the manual masking module 218 via the activation of the manual masking module 218.
  • module 218 will display a set of parameter requests in order to obtain a suitable
  • masking module 218 will relay the obtained masking parameters to the
  • multimedia masking module 214 In a manner similar to the automatic masking
  • multimedia output stream to be presented to the user.
  • identifier 204 could indicate in the captured video stream the beginning and the end of each video segment that shows the customer.
  • module 212 will recognize the requesting user as the specific compliance officer.
  • multimedia stream will include the non-masked image sequences of the video
  • step 302 metadata is extracted from the multimedia input representing
  • the interaction At step 304 the metadata is analyzed in order to determine specific
  • aspects of the multimedia stream generated by the interaction such as the type of
  • the masking method is determined in accordance with events
  • the masking method is received from the masking methods engine.
  • the masking method can be
  • multimedia masking is generated and
  • step 312 selected parts of the multimedia are masked in accordance with the
  • the step of masking can be performed on the fly and when a request for the review of interaction is requested
  • step 314 the selectively masked multimedia is displayed to the
  • step 402 multimedia input is captured by capturing devices.
  • the extracted metadata is
  • a masking method is selected in accordance with the user.
  • the masking method are obtained from the masking method engine and at step 418
  • step 420 the masking elements are inserted into the
  • masking quality could be tuned either
  • the requests steps could utilize an optional "analyze- on-demand"
  • DSP Digital Signal Processing
  • apparatus and method further provides the option of overriding the originally
  • the proposed apparatus could optionally include "smart" triggering
  • the triggering devices could be activate and control recording and storage
  • DSP Digital Signal Processing
  • agents could be any type of resources that can be used in a variety of environments. For example, in call centers, agents could be any type of resources that can be used in a variety of environments. For example, in call centers, agents could be any type of resources that can be used in a variety of environments. For example, in call centers, agents could be any type of resources that can be used in a variety of environments. For example, in call centers, agents could be

Abstract

An apparatus and method for multimedia content based manipulation. The method comprises the masking at least one part of the segment of an interaction, said masking can include the hiding, blurring, or alternatively, the enhancing or focusing on a region of interest.

Description

APPARATUS AND METHOD FOR MULTIMEDIA CONTENT BASED
MANIPULATION
RELATED APPLICATIONS
The present invention relates to US patent application serial number
10/484,107 titled METHOD, APPARATUS AND SYSTEM FOR CAPTURING
AND ANALYZING INTERACTION BASED CONTENT filed July 19, 2001,
and to PCT patent application serial number PCT/IL02/00197 titled A METHOD
FOR CAPTURING, ANALYZING AND RECORDING THE CUSTOMER
SERVICE REPRESENTATIVE ACTIVITIES filed 12 March 2002, and to US
provisional patent application serial number 60/227,478 titled SYSTEM AND
METHOD FOR CAPTURING, ANALYZING AND RECORDING SCREEN
EVENTS filed 24 August 2000 and US patent application serial number
10/056,049 titled VIDEO AND AUDIO CONTENT ANALYSIS filed 30 January
2001, the content of which is hereby incorporated by reference.
BACKGROUND OF THE INVENTION
FIELD OF THE INVENTION
The present invention relates to content derived systems, in general and
to an apparatus and method for manipulating of multimedia based on the content
therein, in particular.
DISCUSSION OF THE RELATED ART
Systems offering capture of interactions comprising voice, video, and
other multimedia, including voice over IP, e-mails, and the like, are currently used by organizations. Organizations capture multimedia interactions in order to
examine the interactions in real time or off line. Reasons for examining
interactions could include the quality assurance of agent's performance, security
purposes, compliance purposes, and the like. For example, call centers use capture
of interaction and logging devices to accomplish the tracking of transactions,
monitor agents' interactions with customers and perform quality assurance on the
interactions performed. Presently access restrictions or limitations to interactions
are based on restrictions based on the identity of the person accessing the
information, the type and identity of the interaction to be accessed or in general
the static properties of the interaction. Thus, a supervisor or an agent wishing to
listen to or view a captured interaction can do so with proper authorization or
access rights. When such an agent or supervisor receives the clearance or has
suitable access rights to an interaction he or she may review, access or manipulate
the entire interaction. In a first example, if the interaction is a telephone call, the
supervisor listening to the interaction can listen to the entire phone call and to all
the speakers participating in the call. In a second example the interaction
comprises a video interaction with associated voice interaction where the captured
voice is substantially synchronized with the video. Currently, the person having
access rights to view the video can view the entire video footage which comprises
the interaction and may listen to the voice coupled there with. Persons requesting
limited access rights in an organization will be not provided with limited or
restricted access to specific aspects or portions of the interaction. A supervisor having access rights to view employee's e-mails will be able to view entire set of
e-mail correspondences despite the fact that some of the e-mails, or messages
coupled within a single e-mail, could be private or could be associated with
another department in the organization. Granted access rights for supervisors to
review interactions performed by agents led call centers to dedicate a telephone
for personal phone; such telephone is not connected to the capture and logging
systems so as to avoid the possibility that a supervisor or another person in the
organization will listen in to personal call. Interactions however, comprise many
aspects and elements which are presently available to such persons receiving
access rights to view the entire interaction. Interactions may comprise a number of
speakers, a plurality of persons appearing within a video footage or stream,
messages received from a number of persons, names of persons carbon copied or
blind carbon copied on messages, data and information which may be related to %
various departments having different access rights or different organizations or
campaigns, and the like.
The presently available systems do not restrict or allow or qualify
access to different aspects or different logical content based segments of the
interactions and does not assign sensitivity level or access rights to elements or
segments of an interaction. A certain segment of the interaction could be vocal,
such as voice spoken by one or more speakers. Such segments of interactions
could also include visual elements, such as the visual presentation of the
appearance of one or more persons (but not all) in a video footage or stream and spoken words by such persons, information or data associated with one issue or
department or campaign or persons within the organization. The access
permission for segments is not time dependent and is not associated with other
activities or with information generated or with policy set in later stages.
Organizations face an ongoing conflict between the requirement to
provide employees at all levels with access to interactions so as to perform their
assigned tasks and the requirement to minimize risks from access to sensitive,
privileged or otherwise inappropriate information. An example that reflects these
requirements can be found in the COPC (Customer Operations Performance
Center) Performance Management System Standards. The COPC Performance
Management System Standards, published during the year 2004, is a set" of
management practices and training for customer-centric service operation that
provides that organizations must have a documented privacy policy that considers
any legal requirements and defines how end-user privacy will be protected. The
COPC guidelines specifically provide that the organization must have rules how
an interaction is monitored and a plan for communicating the findings of all
transactions monitored to staff.
In addition, the cycle of information dissemination is increasingly
shorted requiring on-the-fly solutions for access control of elements or segments
of an interaction. Current systems rely on static definitions for accessing an entire
interaction, at times time dependent, lacking the ability to selectively provide access to elements or segments of an interaction. Automatic or manual access
approvals on-the-fly or even off-line to interactions is not presently available.
There is therefore a need in the art for an apparatus and method to
enable selective and dynamic access to different types, different aspects, and
different segments of an interaction.
SUMMARY OF THE PRESENT INVENTION
One aspect of the present invention regards an apparatus for content based
manipulation of multimedia, such as audio or video. The apparatus comprises a
masking generator device to select a content masking method to be applied to
content, and to activate content masking, and a masking methods engine to
provide a content masking method to the masking generator device.
A second aspect of the present invention regards a method for content
based manipulation of multimedia. The method comprises generating content
masking in accordance with content events, metadata defining content
structure, and content user data.
The advantages of the present invention include the control of access
into segments of an interaction, whether such control includes the masking,
hiding, blurring of information within said interaction. Further such control
can include the restriction to view a particular item, the restriction from
recording a particular interaction or a part thereof, the replacing of a segment
of an interaction with predetermined information, the denial of access to a segment of an interaction or the enhancement of a segment of an interaction.
The control proposed by the present invention is associated with a portion of
the interaction. Each portion of such interaction can be masked or altered or
manipulated in association with the actions and rules associated with the said
interaction and alternatively with the profile of the user accessing the said
interaction.
A third aspect of the present invention includes a method for content
based masking of video, audio or screen streams, said stream comprising one or
more interaction, the method comprising generating content masking on at least a
part of the interaction based on events, or metadata or content user data,
associated with the least one interaction. The method further comprises the step ?
of obtaining metadata associated with the interaction from a content input device.
A content input device can be a camera or a microphone or other data capturing
devices. The method can further comprise the step of masking the metadata
obtained from the content input device and associated with the interaction. A
screen can be any one of the following: a display device, a computer screen, a
video screen, a television screen a portable device screen. The generation of
content masking in association with the interaction can be altered based on one or
more user request. The method can further comprise the step of analyzing the
obtained metadata or the stream associated with the interaction. The method
further comprises the step of associating the metadata associated with the
interaction with a segment of the interaction for which content masking is generated. The method further comprises the step of selecting one or more content
masking to be performed on the one or more interaction to be masked. The
method further comprises the step of altering at least a part of the interaction to be
masked or of inserting a content masking element into the stream associated with
the interaction be masked. The step of generating content masking can comprise
generating content masking from the stream obtained and associated with the
interaction or generating content masking from the stream obtained and associated
with the interaction and subsequently masked interaction. The step of generating
content masking can also comprise generating content masking from a first
masked content and a second masked content; each said masked content is
associated with an interaction. The step of generating content masking can also be
dependant on a previous generation of content masking or on a previously
performed action by the user or an element of the apparatus of the present
invention. The content can be any multimedia content. The user data can be any
multimedia content user data. Masking can include any one of the following:
hiding of information; replacing of information; blurring of images or
information; preventing recording of information; denying access; enhancing a
region of interest; improving visibility of an image or information; encrypting of
information; manipulation of multimedia content based on actions associated with
rules. ■ ' ;
According to a fourth aspect of the present invention there is provided
an apparatus for content based masking of video, audio or screen streams, the stream comprising one or more interaction, the apparatus comprising a content
masking generator for generating content masking on at least a part of the
interaction based on events, or metadata or content user data, associated with the
interaction. The apparatus further comprises a metadata capture device for
obtaining metadata associated with the at least one interaction and one or more
capture device to capture video, audio or screen stream. The apparatus can also
comprise a metadata analysis device for analyzing the metadata obtained from the
metadata capture device and associated with the interaction and a masking
methods device to provide content masking methods to the content masking
generator to be applied to content associated with the interactions. The generation
of content masking associated with the interactions can be altered based on a user
request or other input provided to the apparatus. The apparatus further comprises
an event identification device to recognize events in content associated with the
interaction and based on the results of the analysis of the content, metadata or user
data captured or obtained and a request handler to generate displays and to
respond to commands associated with multimedia display requests, and a masked
streaming device to generate selectively masked content associated with the
interaction. The masking content generators alter at least a part of the interaction
to be masked and inserts content masking elements into the stream associated with
the interaction to be masked during capture or subsequent to the capture, or during
the analysis of the stream or subsequent thereto. The events identification device
identifies single or compound events. The apparatus can further comprise a masked storage to store masked content and an unmasked storage to store
unmasked content. The apparatus further comprises a triggering device to activate
or control content capture, content recording, content storage, or content masking
activation.
BRIEF DESCRIPTION OF THE DRAWINGS
The present invention will be understood and appreciated more fully
from the following detailed description taken in conjunction with the drawings in
which:
Fig. 1 is a schematic block diagram of the multimedia content based
masking apparatus, in accordance with the preferred embodiments of the present
invention;
Fig. 2 is a schematic block diagram of the components of the
multimedia content based masking application, in accordance with the preferred
embodiments of the present invention;
Fig. 3 is a flow chart describing the steps of a method for multimedia
content based masking, in accordance with a preferred embodiment of the present
invention; and
' Fig. 4 is a flow chart describing the steps of a method for multimedia
content based masking, in accordance with another preferred embodiment of the
present invention. DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
An apparatus and method for multimedia content based manipulation is
disclosed. In the context of the present document when the word "masking" is
used it could be construed as one of the meanings listed in the following non-
limiting list: a) hiding, b) replacing, c) blurring, d) preventing recording, e)
denying access, f) enhancing a region of interest, g) improving visibility and the
like manipulation of multimedia content based on actions associated with rules,
and h) transmitting commands to multimedia capture devices, such as camera
Pan-Zoom-Tilt, spectrum change, frame rate change, mute sound, and the like.
In the context of the present invention, an interaction is defined as any
data or information exchanged between at least two individuals or between an.
individual and an information system, such as performing back-office transaction
while interacting via a transactional system. An interaction can comprise one of
the following: a telephone or cellular phone call, a video footage or a stream, an e-
mail, an SMS, a chat session, a web or other network session and the like. While,
referring to multimedia, the preferred embodiment of the present invention refers
to audio, video, screen data and metadata related information interaction or a
combination thereof. Such reference is made without limiting the general
application of the present invention and the scope of the present invention which
will be apparent to persons skilled in the art.
The proposed apparatus and method is designed to enable or restrict
access to interactions according to predetermined organizational access rights based on the content of the interaction or to enhance the multimedia stream
according to and based on dynamic and adaptive manipulation scenarios. The
proposed apparatus and method is further designed to mask, according to its
meanings, segments of portions of interactions according to predetermined
organizational procedures based on the content of the interaction and of the type
of users of the interaction. The masking can be done either in a negative or a
positive manner. Thus, masking can be used to mask the entire content except a
certain portion thereof or could be used to mask only a certain portion of the
content. The masking could be applied to the information stream itself or on the
system metadata. The access to parts of the interaction is allowed or restricted by
using a masking method detailed below based on the content of the interaction, as
well as on other metadata information associated with the interaction. Segments,
portions and elements of the interaction could be enhanced where the
enhancement is based on the content of the interaction, on other metadata
information associated with the interaction, on historical or concurrent
transactions, as well as system requirements like Level of Service (LOS), total
stream size, and the like. The masking is performed based on various
predetermined events and provide predetermined actions for masking (allowing or
restricting multimedia through masking manipulation), also in the layer
configuration of the system of the present invention.
According to the invention content based masking of video, audio or
screen streams comprising one or more interactions are captured and content masking is generated on at least a part of the interaction. The content masking is
based on events, or metadata or content user data, associated with the interactions.
Metadata associated with the interaction is obtained from a content input device
such as a camera or a microphone or other data capturing devices. The generation
of content masking in association with the interaction can be altered based on one
or more user request. Alteration means the changing of the masking applied to the
content associated with the interaction, such as revealing a segment or part
previously hidden and the like. The obtained metadata or the stream associated
with the interaction is further analyzed preferably before masking is performed.
The metadata is associated with the segment of the interaction for which content
masking is generated.
Referring now to Fig.l showing a block diagram of the proposed
apparatus for multimedia content based masking, generally referenced 100. The
apparatus 100 comprises a multimedia capture device 102, a metadata capture
device 104, a metadata analysis device 106, an events identification engine 108, a
mask playback/live request handler 110, a masking generator 112 and a secured
multimedia streamer device 114. The apparatus 100 could also comprise a manual
masking generator 116, a masking methods engine 118, a tuning engine 120, a
content analysis engine 122 and a user presentation generator 124. Optionally, the
apparatus 100 can also comprise one or more data structures to handle the
information associated with the operation of the present invention. Such can
include an events definition data structure 126, an events interaction data structure 128, a user administration data structure 130, and a masking quality data structure
132. In addition, various storage devices can be associated with the apparatus 100.
Such storage devices can further comprise a masked /unmasked multimedia
storage 134, or a general interactions storage device (not shown). Storage devices
can include a disk or a tape device or a DAT device and any other currently
known or later developed storage devices, optical or other which can hold or store
data for request based retrieval. The apparatus 100 is connected or can comprise
one or more multimedia capture devices 102. The multimedia capture devices 102
can include capture devices capturing voice, video, screen events, screens, key
strokes from a keyboard or another device, data streams, e-mail, faxes, chat
sessions, physical access devices and the like. A voice capturing device can be a
telephone or a cellular phone, a satellite phone, a wireless device, a microphone, a
radio frequency receiving device, a voice over IP module or device, a facsimile
machine and the like. A video capture can include the stream captured or the
events occurring within the said video as noted from the changes (events) in the
video stream. The capture device can also include a computing device, mobile or
fixed, in which sessions of data communications occur. Such can include any data
streams or data exchange between individuals or machines, including for example,
e-mail, chat sessions, key strokes on a keyboard, use of virtual keyboards or pens,
■ touch screen use, mouse use, peripherals use and any interaction with a computer.
- Physical access devices can include "smart" access restricting devices or barriers
such as doors which can be opened or locked or controlled from a remote location, cameras or microphone or other recording or logging equipment
controlled from a distance, water and gas mains, electricity supplying stations, and
any other such similar physical device which can be controlled from a distance
and can produce data associated with its physical status. Other such physical
devices can include devices for controlling of public transportation vehicles, such
as subways, trains, elevators, escalators, and any other remotely controlled
aircraft, sea, ground or underground vehicles and the physical devices that control
such vehicles, like crossings, traffic lights, semaphores, landing lights or landing
direction devices, switching devices, and the like. In addition, other remotely
controlled devices can include heating and cooling systems as well as air
conditioning and sewage treatment devices or facilities, In addition, screen events
and screens can be captured as described in detail in U.S. patent application serial
number 10/362,096 titled SYSTEM AND METHOD FOR CAPTURING,
ANALYZING AND RECORDING SCREEN EVENTS filed 24 February 2000
and to US patent application serial number to US patent application serial number
10/362,096 titled SYSTEM AND METHOD FOR CAPTURING BROWSER
SESSIONS AND USER ACTIONS filed 21 February 2003.
In addition, input data from chemical sensors, optical or mechanical
sensors could be captured and processed. Thus data from a chemical sensor could
identify a specific chemical substance or a. specific molecular structure. Data
from optical sensor could provide motion indications and page detections. Mechanical sensors could be associated with scales or various mechanical
counters.
Those skilled in the art will appreciate that any interaction between
individuals or machine can be captured by one or more capture devices and that
the present invention envisage the capturing of such an interaction by one or more
capture devices capturing one or more aspects, such as voice, video, data and the
like, of the interaction. As noted above, the capture devices can be a directly
linked to the apparatus 100 or can be an external capture devices located locally,
or remotely, from which the apparatus 100 receives the captured interactions. The
terms locally can refer to the same room, floor, or building and the term remotely
can refer to another building in the same area or city or even another building or
location, in another city or even country. Locations can include buildings and any
other location where interactions take place, including call centers, banks, offices,
and secured places, public places, such as train stations or airports and the like.
While captured interactions are received from the capture devices enumerated
above, additional information about the multimedia captured can be received from
the metadata associated with the multimedia captured. In association with the
multimedia capture devices the apparatus 100 also comprises a metadata capture
device 102. The metadata received is associated with the multimedia captured and
can include the location of the multimedia capture device 102, or the location of
the interaction, the time the interaction was captured and its length. In addition,
the metadata can include data relating to the multimedia from other systems. Such data can include information stating from which recording systems or logging
systems the capture is received, such as NICE log manufactured by Nice Systems
Ltd. of Israel. In addition, metadata can include environmental conditions,
whether computerized or real world, additional detail as to the scene viewed, the
conversation spoken, the data exchanged, historical statistics, other calls from the
same agent, information on suspicious object descriptors from different
interactions, and the like. One non-limiting example would be information about
the path of an e-mail, a VOIP session, or a chat session. Another non-limiting
example can be information about a person or a location shown within a video
stream captured. Another non-limiting example could include data indicating the
geographical location coordinates of a cellular phone device where the data could,
be obtained from an associated cellular control center. Yet another non-limiting
example could include GPS (Global Positioning System) data indicating'
geographical location coordinates of a satellite phone device.
Persons skilled in the art will appreciate the various other types of
information and data which can be received and processed in association with the
interaction captured. In addition, while the metadata capture device 104 can be
separate to the multimedia capture device 102 it will be easily realized that the
metadata capture device 104 can be a constituent part of the multimedia capture
device 102. Thus, for example, a camera capturing a scene can also transmit the
PTZ parameters used at any given time, the time and date the video stream is
captured, the name of the camera, other environmental parameters, and the like. The metadata capture device 104 may also receive data or information from
telephony related services such as Computer Telephony Interface (CTI), Call
Detail Recording (CDR), Interactive Voice Response (IVR), Dual Tune Multi
Frequency (DTMF), and extension signaling. The metadata analysis device 106
performs an analysis on the metadata received from the metadata capture device
104 and on the interactions captured by the multimedia capture 102. The metadata
analysis 106 uses various analysis on the data received through the use of content
analysis engines 122. The content analysis engines 122 is linked on the one side to
the multimedia capture device 102 and the metadata capture device 104 and
linked on the other side to the metadata analysis device 106. The content analysis
engines 122 could be allocated statically or dynamically in order to optimize
calculation power. The content analysis engines 122 could be available for online
or offline streams for offline availability. The dynamic allocation of the engines
122 is based on metadata capture of offline information, such as, for example,
information indicating an unlawful intrusion into the premises, and the like. The
content analysis engines 122 could include a word spotting content analysis
engine to identify the appearance, or the absence or the appearance of words or a
sequence of words belonging to pre-defined context related lexicon groups to
identify the frequency a word is used or is used in the context of a lexicon; to
identify the frequency words appear which belong to a lexicon group; to identify
words which represent numbers; to identify words that represent a string, a suffix,
a prefix. For example, absence of words can be used to later identify a segment of the application missing which means that a specific interaction cannot be released
outside the organization without counsel's permission for failure to state a
particular disclaimer. The content analysis engines 122 can also include a speech
to text content analysis engine to convert speech into text for processing as text
and to identify to identify full or partial transcription, word count and to obtain the
textual context of the words through identifying whether the context is positive,
negative or belongs to a particular subject. For example, if the context of the
conversation within the interaction in a specific section thereof is negative and the
rest is positive, the later processing in accordance with the present invention may
limit access to the negative context section until approved by a supervisor or a
manager. The content analysis engines 122 can also include a video analysis'"
engine to analyze video streams to identify objects, persons, and object contact,
such as where persons meet or a box or luggage exchanged or where cars collide.
The video analysis engine can also determine the velocity of objects, such as!
whether persons are running or walking or jumping and also recognize faces
through face recognition algorithms. The content analysis engines 122 could also
include a video analysis engine for counting count objects in line or in general,
track suspicious path of objects within a single stream or multiple streams,
detecting abnormal behavior based on comparison with regular behavior detected
in the same scene., and the like The content analysis engines 122 can also include
a screen analysis engine to analyze activities performed on the screen of a
computer through GDI, and data mining analysis to analyze data searches and other data mining activities, CTI analysis to analyze and identify the various CTI
elements, IVR analysis to identify the various IVR elements. In addition, the
content analysis engines 122 can comprise an audio analysis engine to determine
talk over to detect when two or more persons are speaking at the same time,
emotion detection to detect high emotions of one of the speakers, speaker hunting
to detect a particular speaker, language detection to identify the spoken languages,
gender detection to identify the gender of the speakers and silence detection to
detect those parts of the audio interaction which do not include a conversation. An
emotion detector can detect the level of the volume of the speaker's voice and
classify the emotion, such as whether the person is nervous or upset and the like.
In addition, the content analysis engines 122 can also comprise a key stroke
analysis, to determine what was typed, how fast and with what intervals. In '
addition, In addition, the content analysis engines 122 can comprise data analysis '
engine to analyze data received from physical device so as to conclude the '
patterns of use of the physical device. The use of the engines will enable at a later
stage, such as during real-time monitoring, during content playback and during
content retrieval, to restrict access to items analyzed. One example would be to
restrict access to a segment of a video stream shown during a playback operation
initiated by a user that shows two people meeting, or mask a predefined segment
of a video stream showing cars collide having injured people therein, or in the
context of the organization restrict access of supervisors to parts of an interaction
which are private. Likewise, the apparatus of the present invention may restrict access to a video interaction. One non-limiting example relates to a situation in
which certain region covered by a camera is restricted for monitoring, playback,
retrieval or both, due to the presence of a specific object located within the field
of view (FOV) of the camera. Another non-limiting example regards privacy
issues, such as hiding the face of a particular person, such that the face will be
masked, obscured, blurred, enhanced, or the like, throughout the review of the
interaction by a non-approved personal. While such personal may view the video
stream as they may be required to view other elements or parts of the interaction,
the face of a particular person, such as a client may be masked to avoid his or her
identification. In opposite if the face of the client is the item to which access is
granted, the face region of interest can be enhanced while other elements are
blurred or are not enhanced. Another example relates to words uttered or written
in e-mails relating to private matters of the workers in the organization such as
salary or other private information. Once identified and as will be shown here
below the apparatus of the present invention may restrict the access of persons to
the actual confidential details. Thus, a manager, a supervisor or any other
authorized person may be allowed to view a communication involving an
employee, but the employee's communication with human resources or the
employee's salary information will be deleted, hidden, concealed or masked. One
other non-limiting example could be to restrict a sales person from access to parts
of a lead received on a potential sale, if another sales person is already assigned ■
the said lead. An additional non-limiting example concerns a marketing manager or a product manager that are enabled to access only those segments of an
interaction that are directly related to the product line or business details while the
customer may also discuss cross organizational issues. As noted above the
masking operation can be associated with both the hiding or blurring of
information and on the other hand the enhancement of information or a region of
interest. An alternative option would be to mask the entire field of view but for a
specific region of interest. Person skilled in the art will appreciate the many
options available to the user at manipulating the multimedia content in the various
masking methods shown herein above and below.
Still referring to Fig. 1 the events definition data structure 126
comprises a number of data types, such as events, actions and layers. An event is
defined as an occurrence to be identified within the interaction or metadata •
captured. One or more events can comprise a scenario of various activities. For
example, a scenario can be an interaction in which credit card or personal details "
are exchanged. The credit card scenario could involve a set of consecutive events
taking place during the transaction. Such events could include credit card detail
screen capture events, keyboard keystrokes capture events, and the like. Another
example is an interaction in which an agent advises a customer as to possible
investments, and the like. A scenario may comprise one or more events
characterizing the scenario to be identified and marked for later processing. A
plurality of events can characterize a scenario. Events could be identified during
the initial phases of processing or could be based on current data while a retrieval request is submitted. One non-limiting example concerns a security policy
regulated by account identification or by project identification. When a
multimedia content retrieval request is made in order perform content playback
the event will be re-calculated in order to determine whether the requesting party
is a member of the project team or whether he is authorized to deal with the
customer account at the time of the request. The above described logic allows
organizations to enforce a dynamic security policy events and scenarios can be
defined for any type of multimedia. Non-limiting examples, can include be a
certain word spoken, a text written in a written message, an emotion detected, an
object identified, a screen events taken, a data item sent or received, a key stroke
performed and the like. In addition, a defined event can comprise one or more -
events, that is one or more occurrences within the interaction or the metadata can
be predefined to comprise an event. For example, in the credit card details
exchange, one event can be the credit card details screen being loaded by the,
agent to his terminal, key strokes including a template of a credit card, such as
»####_####.####_####•• and in addition identifying that the customer has spoken
numbers during this segment of the interaction. These three events comprise the
credit card details exchange scenario. Alternatively events can be defined as a
single event to be applied to the interactions captured and the associated metadata.
Compound events may be based on the existence or non-existence of "progeny" or
"child" events. One non-limiting example concerns the identification of the
execution of trading actions where a required "progeny" or "child" event, such as the verbal confirmation of the order by the customer, may not exists. In addition
to events the events definition data structure 126 comprises actions. Actions are
defined as the actions to be taken once an event is identified or when the data
retrieval request is submitted. In addition, the data structure 126 comprises a
plurality of layers. Layers are a series of actions associated with an event. A
number of layers may be created for the same event or series of events, thus
providing a number of different actions for the same event or group of events. For
example, if an event is a word spoken, the action in layer 1 can be "mute the
headset for 10 seconds prior to the word". For the same event an action in layer 2
can be "skip 2 seconds from recording before and after the word while streaming
the secured multimedia", or action in layer 3 "replace screen streaming with pre- :
defined image or stream". The data structure 126 also comprises data logical links
associating between events or group of events (such as scenarios) with actions to *
be taken once an event or a group of events is identified and the layers defining ■
which the various actions to be associated with the same event or group of events.
The layer can be associated with the profile of a user. For example, a layer titled
"medical details" can be defined as the actions replacing characters on a screen.
Whenever a non-authorized person access a data structure comprising, among
other data, private medical details, the characters associated with private medical
details are masked on the screen of the viewer. Like examples are personal detail
layers, transaction size layer, and the like, for which different persons in the
organization have access thereto. Persons skilled in the art will appreciate the many other events and scenarios which can be employed in the context of the
present invention. The events identifications engine 108 processes the metadata
received from the metadata analysis device 106 in order to detect the events. The
engine 108 further retrieves the events or group of events to be processed from the
events definition data structure 126 and searches the metadata for the defined
events. The events identifications engine 108 can use a number of alternative
processing modes on the metadata to be processed. The event identifications
module 108 could perform a single pass over the metadata in order to identify one
or more event. If a scenario is searched for the sum of the events of the scenario
will be retrieved and searched on the metadata. The events identification engine
108 module 108 can perform multiple passes over the same metadata in order to
detect compound events.
Still referring to Fig. 1 compound events are events for which a
previous event must be detected to activate the second event. Thus, a first pass is
performed to identify events and then a second pass is performed to identify
events which one condition therein is the existence of the first event. The multiple
passes may be on the same metadata or on metadata associated with different
captured interactions. Similarly, compound events and multiple passes can be
performed. An example of a compound event is an event comprising the existence
of a screen event and the agent sending an e-mail to a particular recipient. In the
first pass the events identifications engine 108 identifies the existence of the
screen event (an event defined in the events definition data structure 126). In the second pass the engine 108 identifies the send e-mail operation and the recipient
name with which the event is associated. It will be realized that during the first
pass the e-mail sending act was not detected as such act is not defined as an event
by itself. Another non-limiting example for a compound event is an event
comprising a credit card data entry and a customer saying the numbers of her
credit card over the telephone while speaking with an agent. The first pass on the
metadata associated with the database of the organization will identify the credit
card data entry and a second pass on the phone interaction with the customer will
identify the audio segment in which the credit card is being provided. When
processing video streams the events identification engine 108 processes each
frame and can also search for predefined vectors. When processing frames, the
engine 108 can identify events within the frames. When the event is identified, the
engine 108 stores the spatial grid location of the event identified. Such can be a
specific location determined according to frame coordinates, and optionally a ;
vector of movement of an object on the screen. The identification of events on the
screen in association with the required actions dictates the data to be later stored
by events identification engine 108 so as to identify the item or object or event
later to be masked. Thus, for example, if an event comprises "locate persons X"
and the metadata associated with the video stream, which is the result of the face
recognition content analysis engine, includes the identity of person X on a number
of frames, the engine 108 will retrieve said frames and use the event associated
action on said frames. If for example, the event action states "determine spatial coordinates for event" the engine 108 will identify the spatial coordinates on each
frame wherein person X appears and will store such parameters in the events
interaction data structure 128 for later use. Alternatively, the action can call for
"identify person X head". In such case the engine 108 retrieves the frames
associated with person X and determines the spatial coordinates and the vector of
"persons x head". Alternatively, the engine 108 can search events or scenarios on
the interactions or metadata on a predetermined or random interval of time. One
non-limiting example can be a search to be performed every 30 seconds or another
like time frame. The engine 108 generates a data record for each metadata for
which events are identified. The data record includes the identification of the
metadata, the identification of the interaction, the location of the event on the
interaction and the actions to be taken when masking is to be performed. Other
data fields may include the associated layer of the actions, the scenarios names,
specific time within the interaction in which the event occurred, the channel
number, the extension number, machine number, the name of the speaker
associated with the event, the area on a frame within which the event was
identified, the vector of movement of the object identified, the data handle and
specific location within said data item identifying the event and the like data
associated with the event, the actions to be taken and the interaction. Many other
like data fields associated with the various events can be stored by engine 108.
The data records associated with the identified events are stored in the events
interactions data structure 128. In addition, the engine 108 can optionally mark the interaction in which an event has been identified. To mark the interaction the
engine 108 stores a status flag on the interaction stream. The status flag is
associated with the data record corresponding to the identification of the event. In
another embodiment the mark is performed by adding a number of data fields to
the interaction, said data fields comprising the data record associated with the data
record to include the event details and the action details.
The mark could be performed by adding a data bit or a number of data
bits at the precise location of the event, said data bits are placed at the start and
end of the event. The data bit comprises a Boolean value. The Boolean value
identifies the start and the end of the event. At the replay of the interaction, the
data bit "alerts" the playback device that an event which may be masked exists
and that a masking event processing should take place prior to playing back the
segment of the interaction marked with the data bits. The data bit can alternatively
comprise the handle for the data record associated with the event and stored in the
events interaction data structure 128. In another alternative the data bit may
include a reference to the layer associated with the events identified so as to allow
the playback device to quickly determine if the person playing back the interaction
is allowed to access the relevant segment marked on the interaction. Alternatively,
the event is noted in association with the interaction, such as on a data structure
associating the interaction and its associated metadata with the event identified,
and not directly and physically thereon so as to maintain the original interaction unchanged. Such data structure includes a table comprising the event handle
associated with an interaction handle and optionally metadata handle. Persons
skilled in the art will appreciate the different methods for noting the event on the
interaction or on the metadata or associated data structures associated with the
event so as to timely identify the event before playback or content retrieval is
performed on the particular segment of the interaction.
Still referring to Fig. I5 the request/push handler 110 is responsible for
handling requests from various users or external applications for access to
interactions either in real-time during on-line monitoring or during off-line
playback and retrieval. The request/push handler 110 will also handle requests
from external applications and/or application integrations utilizing API, Web
services, or any other machine-to-machine communication methods either in a
synchronous or an asynchronous mode. Similarly, handler 110 is responsible for
handling prescheduled push requests to push interactions to recipients. For
example, a supervisor may issue a request through the handler 110 to playback the
recording of an interaction in order to review the interaction which took place
between an agent and a customer. Such request is received and processed by the
request/push handler 110 which retrieves the said request interaction, the events
interaction identified and information from the user administrative data structure
130. The request/push handler 110 can generate metadata that could be used for
masking. To provide a non-limiting example, a video or and audio segment that is
requested more than X times during a Y time interval could trigger a masking rule. The rule will affect the marking of the segment for a manual review by an
officer in order to identify in the content sensitive data or unethical use of the
information. Other requests can concern off-line queries about interactions or
segments thereof, real-time monitoring of an interaction, scenario reconstruction
of an interaction, and the like. User administrative data structure 130 holds
information about the users of the system, including their associated profiles and
security settings or layers. The retrieval request can be for any type of interaction,
including a phone call, a video stream, a screen session, any data or metadata. The
masking generator 112 in association with the masking methods engine 118 is
responsible for the generation of masking method information. The generator 1 12
receives from the request/push handler 110 the data concerning the request and
data regarding the associated user and in accordance with the received
information determines the type, the segment, or the portion of the multimedia
that should be masked. The determination is performed following the operation of
the masking method engine 118 effecting the provision of a suitable multimedia
content masking method. The multimedia content masking method could
comprise video masking, voice masking, data masking, screen masking,
enhancement of multimedia or a region of interest, selective masking or
enhancement of the multimedia content, including of a predetermined region of
interest, and the like. The video masking method could include video block
masking sub-methods and video concealment sub-methods. The video block
masking sub-method could include the following block masking options: a) denial to replay a request, b) implementation of a black screen during play, c)
replacement of the video stream with an alternative stream, such as a stream
containing an informative message, d) a calculation of a motion vector, e) hiding a
block of actions, f) improving the quality of video for masked objects, g) adding
PZT tracking to masked objects, h) processing of multimedia properties related to
masked stream elements, such as gain level, storage options, and the like. The
video concealing sub-method could include the following concealment options: a)
concealing video recorded in the same hour the previous day, b) concealing video
that had been previously authorized for viewing, c) concealing a pre-defined
video, d) deletion of the original multimedia stream based on masking, and the
like. The voice masking method includes a silence implementation sub-method, a
security message or pre-defined audio implementation sub-method, applying
voice transformation method that will allow understanding- of the voice but will
prevent personal identification by voice, and the like. The data masking method
includes a search masking sub-method, a present masking sub-method, and the
like. The search masking sub-method could include mask search of the record,
mask search on a data field, and the like. The present masking sub-method could
include masking the presentation or discovery of the record, masking the
presentation of a data field, and the like. Data masking can be applied on the raw
data level or on the aggregative and descriptive statistical level. Data masking can
also be applied on data transferred to external information systems or databases
such as business intelligence, DWH (Data Warehousing), integration tools, operational or analytical applications such as ERP (Enterprise Resource
Planning), CRM (Customer Relationship Management), PLM (personal Liaison
Management) knowledge management, web portals and any other similar
organizational applications. The screen masking method includes screen block
masking sub-method, a screen concealing sub-method, and the like. The screen
block masking sub-method includes a) denial of the play request, b)
implementation of black screen intervals, c) implementation of replacement
screens, d) hiding a block or a specific field, e) hiding a block specific screen
application, and the like. The screen concealing sub-method includes the insertion
of a stationary field or empty field, the replacement of text with stars, and the like.
Note should be taken than the above described methods, sub-methods
and option could be made effective in various combinations in complex masking
methods. The masking generator 112 relays information concerning the type
and/or the segment to be masked and the information concerning the masking
method to the masking quality data handler 116. The making quality data handler
116 in association with the making quality data structure 132 determines the level
of the masking quality. The masking quality data structure 132 stores masking
quality parameters. Data structure 132 could be tuned optionally by the tuning
engine 120 either automatically or following manual activation. The data handler
116 collects the information concerning all the aspects of the masking and uses
the information to process the multimedia in order to mask specific aspects,
segments, portions, fields, blocks, and the like of the multimedia. The masked multimedia is streamed by the masked multimedia streaming 114 and the masked
multimedia is presented to the user as output 124. Optionally, the masked
multimedia could be stored in the masked multimedia storage 134 for a pre¬
defined period or for a pre-defined volume^ An example of the masking effect is
the blurring of an object or an individual appearing in the masked multimedia.
Masking could be implemented in the following scenario. A service engineer
receives access to an interaction such as a telephone call containing instructions
regarding a required service. The same interaction also includes financial terms
and sums to be paid for the service. The interaction segment associated with the
finances will be blocked for the service engineer's review where various masking
method could be used to accomplish the objective. Thus, the entire segment
containing financial negotiation could be replaced by a replacement screen, such
as a black screen, a static screen, a warning message, or the like, as was described
herein above in association with the masking methods supplied by the masking
methods engine. A masking method could be enhanced via the utilization of a
masking formula (not shown) where the masking formula could be based on
historical event scenarios. The generation of the masking formula could be either
automatic or semi-automatic. As a non-limiting example a video masking process
that may effect the generation or the modification of a masking formula will be
described. In order to provide alarm notifications during monitoring of a video
stream a specific region-of-interest (ROI) is defined on the video images
constituting the video stream. The ROI is defined on the video stream by the utilization of Field of View (FOV) limitation. A pre-defined event taking place
inside the defined region-of-interest effect the activation of an alarm. If
historically the group of total alarms activated includes a proportionally high
number of false alarms then one of the solutions is to reduce the size of the ROI.
The reduction could be made semi-automatically by recommending to the user an
action for the reduction of the ROI zone. Alternatively, the ROI reduction could
be accomplished automatically via the generation of the mask formula.
Consequently, the mask formula will include an improved masking algorithm that
will reduce sensitivity in the area wherein the high number of false alarms was
generated. In yet another non-limiting example the mask formula could be used
improving a masking algorithm where historically a high number of alarms are
generated where the alarms indicate the false identification of a specifically
defined word. The problem could be solved either semi-automatically or
automatically. In the semi-automatic mode the user could be advised to activate an
improved algorithm when the word is sensed. In the automatic mode a masking
formula including the masking algorithm and associated with the masking process
could be generated or modified in order to provide an improved algorithm when
the problematic word is sensed. The masking formula could be also used
implementing general masking options. For example, the masking formula could
include data indicating interaction types, interaction segments, and interaction
segments that should be always masked or alternatively could include data
indicating interaction types, interaction segments, and interaction segments that are not to be masked. Multimedia elements to be masked in accordance with the
masking formula could include entire management interactions, (phone calls with
managers as participants), specific areas in human body (based on body
proportions), pre-defined types of security personnel (police, bodyguards, FBI
employees, CIA employees, and the like).
Note should be taken that although the drawing under discussion
shows a single multimedia capture device, a single metadata capture device, and
a set of single devices, it would be readily perceived that in a realistic
environment a multitude of capture could be sending a plurality of multimedia
streams and metadata information to a plurality of metadata analyzers, events
identification engines, and request handlers., Further note should be taken that the
apparatus presented is exemplary only. In other preferred embodiments of the
present invention, the entire set of devices, modules, engines, data structures, and
handlers could be co-located on the same computing platform. In yet further
embodiments of the present invention, a multiplexing device could be added in
order to multiplex several multimedia streams and several metadata records from
several capturing devices into a multiplexed and combined multimedia and
metadata record stream. A de-multiplexer unit could be placed between the
capture devices and the metadata analysis device in order to separate the
combined multimedia/metadata stream prior to the suitable processing.
Referring now to Fig. 2 an apparatus 200 is shown in accordance with
a preferred embodiment of the present invention. The apparatus 200 includes an interaction handler 202, an events definition handler and identifier 204, a
request/push handler 206, a masking methods selector module 208, a masking
quality data handler 210, a masking generator module 212, a multimedia masking
module 214, a manual masking module 218, and a secured (masked output)
handler 216. The interaction handler 202 manages the multimedia interactions
performed by the interaction participants, such two or more individuals. The
interaction could be performed directly face-to-face or could be performed via
diverse communication channels, such as a telephone line, computerized e-mail,
fax exchanges, Web-based chat sessions and the like. The management of the
interaction involves the capturing of the interaction data via diverse multimedia
capturing devices corresponding to the interaction type and the interaction
channels. The interaction management further includes the capturing of the
interaction metadata or the extraction of the metadata from the multimedia. The
metadata includes control information embedded in the interaction data. Thus,
metadata could include indications concerning DTMF signaling, e-mail addresses,
interaction capturing device types and device identifications, the capture of certain
keystrokes on an input keyboard, the activation of function keys associated with a
computing device, the pointing and clicking of a pointer device, the activation of
keystrokes on a cellular phone signifying the initiation of an SMS message, and
the like. In accordance with the interaction channels and the interaction media the
metadata could be captured separately or could be extracted from the interaction
content. The captured metadata could include locations, time and date, participants' identification, priority parameters, CDR/CTI information IVR
information, DTMF, extension signaling, camera parameters, environmental
parameters, and date related from external systems. Typically, the interaction
content is captured in a multimedia format, such as video, audio, graphics, text,
and the various combinations thereof. The metadata associated with the
multimedia content is analyzed in order to identify the control information carried
thereby. The metadata analysis could include diverse analysis techniques, such as
word spotting, speech-to-text, emotion detection, video analysis, screen analysis,
audio analysis, data mining, and keystroke analysis. The multimedia content is
then streamed to the event definition handler and identifier 204. The handler and
identifier 204 processes the multimedia stream in order to identify, to recognize
and to classify the various events embedded in the multimedia for later handling.
The requests/push handler 206 is responsible for both the automatic initiation of
multimedia replay or retrieval requests and for the generation of multimedia
content replay requests in response to manual replay commands submitted by the
users of the apparatus. The automatic or user command-generated requests
activate the masking generator module 212. The generator 212 collects
multimedia content masking related information, and activates the masking
method selector module 208. The masking method selector module 208
determines the suitable multimedia content based masking method that is to be
applied to the multimedia stream to be replayed in accordance with the
automatically generated or manual command generated replay requests. The masking generator 212, in association with the masking quality data handler 210,
relays the masking information to the multimedia masking module 214. The
secured (masked) output handler 216 applies the masking method to the
appropriate types, aspects, segments, screens, blocks, motion vectors, or fields of
the multimedia stream in order to conceal specifically pre-defined information
from the user. Optionally, the masking could be performed manually by the user
via the activation of the manual masking module 218. The manual masking
module 218 will display a set of parameter requests in order to obtain a suitable
set of relevant masking parameters from the user. Consequently, the manual
masking module 218 will relay the obtained masking parameters to the
multimedia masking module 214. In a manner similar to the automatic masking
process the multimedia masking module 214 will relay the masking parameters to
the secured (masked) output handler 216 that will generate a suitably masked
multimedia output stream to be presented to the user.
Still referring to Fig. 2 in an exemplary interaction scenario a face-to-
face interaction is performed between a customer and an agent. It could have been
pre-defined that as a result of privacy considerations a certain compliance officer
where monitoring the interaction is not allowed to see the video of the customer.
As a result the video stream generated by the video capture device covering the
customer should be blacked out or replaced with a replacement image during a
replay requested by the same compliance officer. The events definition handler
and identifier 204 could indicate in the captured video stream the beginning and the end of each video segment that shows the customer. The masking generator
module 212 will recognize the requesting user as the specific compliance officer.
Thus, a specific masking method will be selected that will effect the blocking out
of all the customer-specific video segments. During playback the display of the
multimedia stream will include the non-masked image sequences of the video
segments in which the agent appears, the non-masked audio stream of both the
agent and the customer, and a blacked out video segment with an optional
notification to the viewer replacing the image sequences of the customer.
Referring now to Fig. 3 in the preferred embodiment of the present
invention at step 302 metadata is extracted from the multimedia input representing
the interaction. At step 304 the metadata is analyzed in order to determine specific
aspects of the multimedia stream generated by the interaction, such as the type of
the interaction, the origin of the interaction and other metadata as detailed above.
At step 306 the masking method is determined in accordance with events
definitions and optionally also with associated user data. At step 308 a masking
method is received from the masking methods engine. The masking method can
be any form of masking of the data, whether through showing details or hiding
details as is described in detail in association with Fig. 1. The masking method to
be used can be predetermined. At step 310 multimedia masking is generated and
at step 312 selected parts of the multimedia are masked in accordance with the
masking method by the processing of the suitable part of the multimedia or by the
insertion of the masking elements into the multimedia. The step of masking can be performed on the fly and when a request for the review of interaction is requested
or in advance on raw data to be saved as masked data to be later retrieved by the
requesting user. At step 314 the selectively masked multimedia is displayed to the
user.
Referring now to Fig. 4 in another preferred embodiment of the present
invention at step 402 multimedia input is captured by capturing devices. At step
404 metadata indicating specific aspects of the captured multimedia stream are
extracted from the multimedia stream. At step 406 the extracted metadata is
analyzed and at step 408 event definitions and interactions are obtained from the
events definitions and interactions data structure. At step 410 retrieval requests
are generated automatically or in response to user submitted commands. At step
412 information regarding the user is obtained from a data structure storing user
data. At step 414 a masking method is selected in accordance with the user. data,
the events definitions, and the analyzed metadata. At step 416 the parameters of
the masking method are obtained from the masking method engine and at step 418
automatic multimedia masking is generated. Optionally, the multimedia masking
could be generated in response to the manual insertion of masking parameters by
the user (step 420). At step 422 the masking elements are inserted into the
multimedia stream and at step 424 the selectively masked multimedia is displayed
to the user. Optionally, at step 426 masking quality could be tuned either
automatically or in response to the manual insertion of suitable masking quality
parameters by the user. Notes should be taken that in other preferred embodiments of the
invention the sequence of the above described steps could differ. As a non-
limiting example, the steps 410 through 412 that deal with the retrieval requests
could be performed prior to the steps 406 through 408 that deal with events
identification. The requests steps could utilize an optional "analyze- on-demand"
parameter that could be set either automatically or could be obtained in real-time
from a user submitting the request. A specific pre-determined value of the
"analyze-on-demand" parameter could indicate an "analyze" command that will
effect the performance of the event identification steps (406 through 408) and the
subsequent masking, steps (414 through 426) after the retrieval steps (410 through
412). An another pre-determined value of the "analyze-on-demand" parameter
could indicate a "do not analyze command" and will effect the display of the
requested multimedia only while the analysis, event identification and masking
steps will be bypassed..
Further note should be taken that both the enhancement and the
masking options could be implemented in real-time and nearly real-time via the
utilization of Digital Signal Processing (DSP) devices and methods. The proposed
apparatus and method further provides the option of overriding the originally
recorded multimedia stream on the recording platform and on the storage level.
For example, a video recording for the monitoring in a privacy-sensitive
environment, such as a public restroom, the field of view (FOV) of the captured
image could be limited. The proposed apparatus could optionally include "smart" triggering
devices on each level of the architecture and in association with each component.
The triggering devices could be activate and control recording and storage
processes and options, DSP (Digital Signal Processing) processes and options,
online and off-line (batch) processes, user input, and the like. For example, the
type of the display device used for monitoring, playback or retrieval could trigger
content masking, such as blocking and concealing or content enhancing. Thus,
when a mobile device is used as the display device the type of the device effect
either more rigorous masking or high resolution enhancing in order to provide for
better tracking or enhanced access control for a specific object, such as
individuals and other objects in motion.
The apparatus and method proposed by the present invention could be
utilized in a variety of environments. For example, in call centers, agents could be
provided with the option to mark certain calls as private and thereby could prevent
monitoring of the call. In healthcare industries the proposed apparatus and method
could be used for concealing medical personal information from un-authorized
personnel, or for tracing all access requests to specific medical information. In
call centers and banking institutions the proposed apparatus and method could be
utilized for protecting personal information and for concealing system credentials.
In government public safety applications selected parts of recordings submitted to
courts could be masked for legal reasons, such as court restrictions on revealing
the identify or witnesses. It will be appreciated by persons skilled in the art that the present
invention is not limited to what has been particularly shown and described
hereinabove. Rather the scope of the present invention is defined only by the
claims which follow.

Claims

CLAIMSI/WE CLAIM:
1. A method for content based masking of video, audio or screen streams,
said stream comprising at least one interaction, the method comprising
generating content masking on at least a part of the at least one interaction
based on events, or metadata or content user data, associated with the least
one interaction.
2. The method of claim 1 further comprises the step of obtaining metadata
associated with the at least one interaction from a content input device.
3. The method of claim 2 further comprises the step of masking the metadata
obtained from the content input device and associated with the at least one
interaction.
4. The method of claim 1 wherein the screen comprises any one of the
following: a display device, a computer screen, a video screen, a television
screen a portable device screen.
5. The method for claim 1 wherein the generation of content masking of the
at least one interaction is altered based on at least one user request.
6. The method for claim 1 further comprises the step of analyzing the
obtained metadata or the stream associated with the at least one interaction.
7. The method for claim 1 further comprising the step of associating the
metadata associated with the at least one interaction with a segment of the
interaction for which content masking is generated.
8. The method of claim 1 further comprises the step of selecting an at least
one content masking to be performed on the at least one interaction to be
masked.
9. The method of claim 1 further comprises the step of altering at least a part
of the interaction to be masked.
10. The method of claim 1 further comprises the step of inserting an at least
one content masking element into the stream associated with the at least
one interaction be masked.
11. The method of claim 1 wherein the step of generating content masking
comprises generating content masking from the stream obtained and
associated with the at least one interaction.
12. The method of claim 1 wherein the step of generating content masking
comprises generating content masking from the stream obtained and
associated with the at least one interaction and subsequently masked.
13. The method of claim 1 wherein the step of generating content masking
comprises generating content masking from a first masked content and a
second masked content, each said masked content is associated with an at
least one interaction.
14. The method of claim 1 wherein the step of generating content masking is
associated on at least one previous generation of content masking or on a
previously performed action by the user or an element of the apparatus of
the present invention.
15. The method of claim 1 wherein the content is multimedia content.
16. The method of claim 1 wherein the content user data is multimedia content
user data.
17. The method of claim 1 wherein masking is one of the following: hiding of
information; replacing of information; blurring of images or information;
preventing recording of information; denying access; enhancing a region of
interest; improving visibility of an image or information; encrypting of
information; manipulation of multimedia content based on actions
associated with rules.
18. An apparatus for content based masking of video, audio or screen streams,
said stream comprising at least one interaction, the apparatus comprising a
content masking generator for generating content masking on at least a part
of the at least one interaction based on events, or metadata or content user
data, associated with the at least one interaction.
19. The apparatus of claim 18 further comprises a metadata capture device for
obtaining metadata associated with the at least one interaction.
20. The apparatus of claim 18 further comprises at least one capture device to
capture video, audio or screen stream.
21. The apparatus of claim 19 further comprises a metadata analysis device
for analyzing the metadata obtained from the metadata capture device and
associated with the at least one interaction.
22. The apparatus of claim 18 wherein the screen comprises any one of the
following: a display device, a computer screen, a video screen, a television
screen a portable device screen.
23. The apparatus for claim 18 further comprising a masking methods device
to provide an at least one content masking method to the content masking
generator to be applied to content associated with the at least one
interaction.
24. The apparatus of claim 18 wherein the generation of content masking of
the at least one interaction is altered based on at least one user request.
25. The apparatus of claim 18 further comprises an event identification device
to recognize events in content associated with the at least one interaction
and based on the results of the analysis of the content, metadata or user
data captured or obtained.
26. The apparatus of claim 18 further comprises a request handler to generate
displays and to respond to commands associated with multimedia display
requests, and a masked streaming device to generate selectively masked
content associated with the at least one interaction.
27. The apparatus of claim 18 wherein the masking content generator alters at
least a part of the interaction to be masked.
28. The apparatus of claim 18 wherein the masking content generator inserts
an at least one content masking element into the stream associated with the
at least one interaction to be masked.
29. The apparatus of claim 18 wherein the content to be masked is multimedia
content.
30. The apparatus of claim 18 wherein the content user data is multimedia
content user data.
31. The apparatus of claim 18 wherein masking is one of the following: hiding
of information; replacing of information; blurring of images or
information; preventing recording of information; denying access;
enhancing a region of interest; improving visibility of an image or
information; encrypting of information; manipulation of multimedia
content based on actions associated with rules.
32. The apparatus of claim 25 wherein the events identification device
identifies single events.
33. The apparatus of claim 25 wherein the event identification identifies
recognizes compound events.
34. The apparatus of claim 18 further comprises a masked storage to store
selectively masked content.
35. The apparatus of claim 18 further comprises an unmasked storage to store
unmasked content.
36. The apparatus of claim 18 further comprises at least one triggering device
to activate or control content capture, content recording, content storage, or
content masking activation.
PCT/IL2004/000736 2004-08-09 2004-08-09 Apparatus and method for multimedia content based WO2006021943A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US10/507,582 US7714878B2 (en) 2004-08-09 2004-08-09 Apparatus and method for multimedia content based manipulation
PCT/IL2004/000736 WO2006021943A1 (en) 2004-08-09 2004-08-09 Apparatus and method for multimedia content based

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/IL2004/000736 WO2006021943A1 (en) 2004-08-09 2004-08-09 Apparatus and method for multimedia content based

Publications (1)

Publication Number Publication Date
WO2006021943A1 true WO2006021943A1 (en) 2006-03-02

Family

ID=35967195

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IL2004/000736 WO2006021943A1 (en) 2004-08-09 2004-08-09 Apparatus and method for multimedia content based

Country Status (2)

Country Link
US (1) US7714878B2 (en)
WO (1) WO2006021943A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7587454B2 (en) 2004-12-08 2009-09-08 Nice Systems Ltd. Video streaming parameter optimization and QoS
WO2011045116A1 (en) * 2009-10-15 2011-04-21 Thomson Licensing Method for adding voice content to video content and device for implementing said method
WO2016012484A1 (en) * 2014-07-23 2016-01-28 Koninklijke Philips N.V. Device and method for de-identification of information to be displayed
WO2016042190A1 (en) * 2014-09-15 2016-03-24 Tecnología Sostenible Y Responsable Sl Method and portable electronic device having a double screen
CN105939334A (en) * 2015-03-04 2016-09-14 费希尔-罗斯蒙特系统公司 Anomaly detection in industrial communications networks
WO2020089917A1 (en) * 2018-11-02 2020-05-07 BriefCam Ltd. Method and system for automatic object-aware video or audio redaction

Families Citing this family (94)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005086620A2 (en) * 2003-10-10 2005-09-22 L-3 Communications Security And Detection Systems Mmw contraband screening system
US8276096B2 (en) * 2004-04-02 2012-09-25 International Business Machines Corporation Multicast file viewing and editing
WO2006021943A1 (en) * 2004-08-09 2006-03-02 Nice Systems Ltd. Apparatus and method for multimedia content based
US7650628B2 (en) * 2004-10-21 2010-01-19 Escription, Inc. Transcription data security
JP4312246B2 (en) * 2005-03-18 2009-08-12 富士通株式会社 Moving image data editing apparatus, moving image data reproducing apparatus, moving image data editing method, and moving image data reproducing method
US8094803B2 (en) * 2005-05-18 2012-01-10 Mattersight Corporation Method and system for analyzing separated voice data of a telephonic communication between a customer and a contact center by applying a psychological behavioral model thereto
US8094790B2 (en) 2005-05-18 2012-01-10 Mattersight Corporation Method and software for training a customer service representative by analysis of a telephonic interaction between a customer and a contact center
US20070028262A1 (en) * 2005-07-29 2007-02-01 Zermatt Systems, Inc. Virtual discovery of content available to a device
JP2007041988A (en) * 2005-08-05 2007-02-15 Sony Corp Information processing device, method and program
WO2007111648A2 (en) * 2005-10-28 2007-10-04 Openconnect Systems, Incorporated Modeling interactions with a computer system
US20070116328A1 (en) * 2005-11-23 2007-05-24 Sezai Sablak Nudity mask for use in displaying video camera images
US7562822B1 (en) * 2005-12-30 2009-07-21 Leapfrog Enterprises, Inc. Methods and devices for creating and processing content
US20070244925A1 (en) * 2006-04-12 2007-10-18 Jean-Francois Albouze Intelligent image searching
US20070260627A1 (en) * 2006-05-03 2007-11-08 Lucent Technologies Inc. Method and apparatus for selective content modification within a content complex
JP4175390B2 (en) * 2006-06-09 2008-11-05 ソニー株式会社 Information processing apparatus, information processing method, and computer program
US8655916B2 (en) * 2006-09-22 2014-02-18 Yahoo! Inc. System and method for creating user profiles
WO2008057285A2 (en) * 2006-10-27 2008-05-15 Vidient Systems, Inc. An apparatus for image capture with automatic and manual field of interest processing with a multi-resolution camera
KR100828371B1 (en) * 2006-10-27 2008-05-08 삼성전자주식회사 Method and Apparatus of generating meta data of content
US8161395B2 (en) * 2006-11-13 2012-04-17 Cisco Technology, Inc. Method for secure data entry in an application
US8718262B2 (en) 2007-03-30 2014-05-06 Mattersight Corporation Method and system for automatically routing a telephonic communication base on analytic attributes associated with prior telephonic communication
US20080240404A1 (en) * 2007-03-30 2008-10-02 Kelly Conway Method and system for aggregating and analyzing data relating to an interaction between a customer and a contact center agent
US8023639B2 (en) 2007-03-30 2011-09-20 Mattersight Corporation Method and system determining the complexity of a telephonic communication received by a contact center
US20080240374A1 (en) * 2007-03-30 2008-10-02 Kelly Conway Method and system for linking customer conversation channels
FR2917931A1 (en) * 2007-06-22 2008-12-26 France Telecom METHOD AND SYSTEM FOR CONNECTING PEOPLE IN A TELECOMMUNICATIONS SYSTEM.
US20090019553A1 (en) * 2007-07-10 2009-01-15 International Business Machines Corporation Tagging private sections in text, audio, and video media
US8224087B2 (en) * 2007-07-16 2012-07-17 Michael Bronstein Method and apparatus for video digest generation
US8112715B2 (en) * 2007-07-31 2012-02-07 International Business Machines Corporation Content management system that renders a document to a user based on a usage profile that indicates previous activity in accessing the document
US10419611B2 (en) 2007-09-28 2019-09-17 Mattersight Corporation System and methods for determining trends in electronic communications
US8624733B2 (en) * 2007-11-05 2014-01-07 Francis John Cusack, JR. Device for electronic access control with integrated surveillance
US9270950B2 (en) 2008-01-03 2016-02-23 International Business Machines Corporation Identifying a locale for controlling capture of data by a digital life recorder based on location
US9164995B2 (en) 2008-01-03 2015-10-20 International Business Machines Corporation Establishing usage policies for recorded events in digital life recording
US9105298B2 (en) * 2008-01-03 2015-08-11 International Business Machines Corporation Digital life recorder with selective playback of digital video
US9769542B2 (en) * 2008-03-28 2017-09-19 Verint Americas Inc. System and method for digital rights management control using video analytics
GB0808448D0 (en) * 2008-05-09 2008-06-18 Elync Ltd Secure communication system and method of operating the same
US8204180B1 (en) * 2008-08-08 2012-06-19 Intervoice Limited Partnership Systems and methods for preventing sensitive information from being communicated into a non-secure environment
US20100070437A1 (en) * 2008-09-15 2010-03-18 International Business Machines Corporation Information Management for Information Display Systems
EP2338278B1 (en) 2008-09-16 2015-02-25 Intel Corporation Method for presenting an interactive video/multimedia application using content-aware metadata
US8868661B2 (en) * 2008-10-08 2014-10-21 Verizon Patent And Licensing Inc. Message management based on metadata
JP2012513146A (en) * 2008-12-19 2012-06-07 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Control display parameter settings
EP2247077A1 (en) * 2009-04-03 2010-11-03 Hewlett-Packard Development Company, L.P. Method and apparatus for network communications
WO2011004499A1 (en) * 2009-07-10 2011-01-13 富士通株式会社 Electronic device, security method therefor, security program therefor, and recording medium
US8405499B2 (en) 2009-08-07 2013-03-26 At&T Intellectual Property I, L.P. Methods, systems, and products for security services
US8937658B2 (en) 2009-10-15 2015-01-20 At&T Intellectual Property I, L.P. Methods, systems, and products for security services
US20110191683A1 (en) * 2009-12-04 2011-08-04 Dillard Daniel G Methods and Systems to Enhance Advisor-Client Communications
US20110218798A1 (en) * 2010-03-05 2011-09-08 Nexdia Inc. Obfuscating sensitive content in audio sources
GB2478916B (en) * 2010-03-22 2014-06-11 Veritape Ltd Transaction security method and system
CN103069457A (en) * 2010-08-10 2013-04-24 Lg电子株式会社 Region of interest based video synopsis
US10908794B2 (en) * 2010-08-16 2021-02-02 Iheartmedia Management Services, Inc. Automated scheduling of multimedia content avoiding adjacency conflicts
US8694667B2 (en) 2011-01-05 2014-04-08 International Business Machines Corporation Video data filtering method and system
WO2012149448A1 (en) * 2011-04-29 2012-11-01 Wellness & Prevension, Inc. Method and system for user initiated electronic messaging
US8832077B1 (en) 2011-05-04 2014-09-09 Google Inc. Selectively retrieving search results in accordance with different logical relationships
JP5871485B2 (en) * 2011-05-17 2016-03-01 キヤノン株式会社 Image transmission apparatus, image transmission method, and program
US9026596B2 (en) * 2011-06-16 2015-05-05 Microsoft Technology Licensing, Llc Sharing of event media streams
US9111239B2 (en) * 2011-07-13 2015-08-18 Incontact, Inc. Activity recording of contact handling system agents
US9659266B2 (en) 2011-07-14 2017-05-23 International Business Machines Corporation Enterprise intelligence (‘EI’) management in an EI framework
US8566345B2 (en) 2011-07-14 2013-10-22 International Business Machines Corporation Enterprise intelligence (‘EI’) reporting in an EI framework
US9646278B2 (en) 2011-07-14 2017-05-09 International Business Machines Corporation Decomposing a process model in an enterprise intelligence (‘EI’) framework
US9639815B2 (en) 2011-07-14 2017-05-02 International Business Machines Corporation Managing processes in an enterprise intelligence (‘EI’) assembly of an EI framework
US9965564B2 (en) 2011-07-26 2018-05-08 Schneider Electric It Corporation Apparatus and method of displaying hardware status using augmented reality
US9110933B1 (en) * 2011-11-04 2015-08-18 Google Inc. Processing data triggers in an untrusted environment based on information stored in a trusted environment
US9396634B2 (en) 2011-11-10 2016-07-19 At&T Intellectual Property I, L.P. Methods, systems, and products for security services
US9379915B2 (en) 2011-11-10 2016-06-28 At&T Intellectual Property I, L.P. Methods, systems, and products for security services
US8902740B2 (en) * 2011-11-10 2014-12-02 At&T Intellectual Property I, L.P. Methods, systems, and products for security services
US9277381B2 (en) 2011-11-10 2016-03-01 At&T Intellectual Property I, L.P. Methods, systems, and products for security services
US8692665B2 (en) 2011-11-10 2014-04-08 At&T Intellectual Property I, L.P. Methods, systems, and products for security services
US9064013B1 (en) 2011-11-30 2015-06-23 Google Inc. Application of resource limits to request processing
US9235607B1 (en) 2012-03-29 2016-01-12 Google Inc. Specifying a predetermined degree of inconsistency for test data
WO2014028038A1 (en) * 2012-08-15 2014-02-20 Hewlett-Packard Development Company, Lp Validating a metadata tree using a metadata integrity validator
CN103310411B (en) * 2012-09-25 2017-04-12 中兴通讯股份有限公司 Image local reinforcement method and device
JP2014137627A (en) * 2013-01-15 2014-07-28 Sony Corp Input apparatus, output apparatus, and storage medium
US10346621B2 (en) * 2013-05-23 2019-07-09 yTrre, Inc. End-to-end situation aware operations solution for customer experience centric businesses
US20150085146A1 (en) * 2013-09-23 2015-03-26 Nvidia Corporation Method and system for storing contact information in an image using a mobile device
US10410648B1 (en) * 2013-12-31 2019-09-10 Allscripts Software, Llc Moderating system response using stress content of voice command
KR102308763B1 (en) * 2015-01-20 2021-10-05 삼성전자주식회사 Apparatus and method for tightening security of personal information data
EP3262833B1 (en) 2015-02-24 2023-02-22 Axon Enterprise, Inc. Systems and methods for bulk redaction of recorded data
TW201705058A (en) * 2015-07-28 2017-02-01 廣達電腦股份有限公司 Information push system and method
US10373453B2 (en) 2015-09-15 2019-08-06 At&T Intellectual Property I, L.P. Methods, systems, and products for security services
US10565840B2 (en) 2015-11-12 2020-02-18 At&T Intellectual Property I, L.P. Alarm reporting
US9460168B1 (en) * 2016-02-17 2016-10-04 Synclayer, LLC Event visualization
US9965841B2 (en) 2016-02-29 2018-05-08 Schneider Electric USA, Inc. Monitoring system based on image analysis of photos
US10776519B2 (en) * 2016-08-08 2020-09-15 International Business Machines Corporation Information presentation management
EP3526964B1 (en) 2016-10-14 2024-02-21 Genetec Inc. Masking in video stream
US10902147B2 (en) * 2016-11-04 2021-01-26 Intellisist, Inc. System and method for performing screen capture-based sensitive information protection within a call center environment
US20180262404A1 (en) * 2017-03-13 2018-09-13 Microsoft Technology Licensing, Llc Hypermedia-driven record and playback test framework
US10909978B2 (en) * 2017-06-28 2021-02-02 Amazon Technologies, Inc. Secure utterance storage
US10846429B2 (en) 2017-07-20 2020-11-24 Nuance Communications, Inc. Automated obscuring system and method
US11222081B2 (en) 2017-11-27 2022-01-11 Evoqua Water Technologies Llc Off-line electronic documentation solutions
JP2020052145A (en) * 2018-09-25 2020-04-02 トヨタ自動車株式会社 Voice recognition device, voice recognition method and voice recognition program
US10762607B2 (en) 2019-04-10 2020-09-01 Alibaba Group Holding Limited Method and device for sensitive data masking based on image recognition
US10719966B1 (en) 2019-06-11 2020-07-21 Allstate Insurance Company Accident re-creation using augmented reality
US11620361B2 (en) 2021-03-03 2023-04-04 International Business Machines Corporation Proactive privacy content hosting
US11330229B1 (en) * 2021-09-28 2022-05-10 Atlassian Pty Ltd. Apparatuses, computer-implemented methods, and computer program products for generating a collaborative contextual summary interface in association with an audio-video conferencing interface service
WO2023069456A2 (en) * 2021-10-18 2023-04-27 Apprentice FS, Inc. Method for distributing censored videos of manufacturing procedures performed within a facility to remote viewers
US11722536B2 (en) 2021-12-27 2023-08-08 Atlassian Pty Ltd. Apparatuses, computer-implemented methods, and computer program products for managing a shared dynamic collaborative presentation progression interface in association with an audio-video conferencing interface service

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040267952A1 (en) * 2003-06-24 2004-12-30 He Li-Wei Variable play speed control for media streams
US6843110B2 (en) * 2002-06-25 2005-01-18 Fluid Components International Llc Method and apparatus for validating the accuracy of a flowmeter
US20050033760A1 (en) * 1998-09-01 2005-02-10 Charles Fuller Embedded metadata engines in digital capture devices

Family Cites Families (110)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4145715A (en) * 1976-12-22 1979-03-20 Electronic Management Support, Inc. Surveillance system
US4493027A (en) * 1981-05-22 1985-01-08 Data General Corporation Method of performing a call operation in a digital data processing system having microcode call and return operations
US4525780A (en) * 1981-05-22 1985-06-25 Data General Corporation Data processing system having a memory using object-based information and a protection scheme for determining access rights to such information
US4527151A (en) * 1982-05-03 1985-07-02 Sri International Method and apparatus for intrusion detection
US4821118A (en) * 1986-10-09 1989-04-11 Advanced Identification Systems, Inc. Video image system for personal identification
US4930160A (en) * 1987-09-02 1990-05-29 Vogel Peter S Automatic censorship of video programs
US5313227A (en) * 1988-04-15 1994-05-17 International Business Machines Corporation Graphic display system capable of cutting out partial images
US5353618A (en) * 1989-08-24 1994-10-11 Armco Steel Company, L.P. Apparatus and method for forming a tubular frame member
GB9000105D0 (en) 1990-01-03 1990-03-07 Racal Recorders Ltd Recording system
US5051827A (en) * 1990-01-29 1991-09-24 The Grass Valley Group, Inc. Television signal encoder/decoder configuration control
US5091780A (en) * 1990-05-09 1992-02-25 Carnegie-Mellon University A trainable security system emthod for the same
EP0484076B1 (en) * 1990-10-29 1996-12-18 Kabushiki Kaisha Toshiba Video camera having focusing and image-processing function
DE69124777T2 (en) * 1990-11-30 1997-06-26 Canon Kk Device for the detection of the motion vector
GB2259212B (en) * 1991-08-27 1995-03-29 Sony Broadcast & Communication Standards conversion of digital video signals
GB2268354B (en) * 1992-06-25 1995-10-25 Sony Broadcast & Communication Time base conversion
US8272958B2 (en) * 2004-01-26 2012-09-25 Shuffle Master, Inc. Automated multiplayer game table with unique image feed of dealer
US5369739A (en) * 1993-07-09 1994-11-29 Silicon Graphics, Inc. Apparatus and method for generating point sample masks in a graphics display system
US5519446A (en) * 1993-11-13 1996-05-21 Goldstar Co., Ltd. Apparatus and method for converting an HDTV signal to a non-HDTV signal
US5491511A (en) * 1994-02-04 1996-02-13 Odle; James A. Multimedia capture and audit system for a video surveillance network
IL113434A0 (en) * 1994-04-25 1995-07-31 Katz Barry Surveillance system and method for asynchronously recording digital data with respect to video data
US5687303A (en) * 1994-05-18 1997-11-11 Xerox Corporation Printer controller for object optimized printing
US6327043B1 (en) * 1994-05-18 2001-12-04 Xerox Corporation Object optimized printing system and method
US5555441A (en) * 1994-08-02 1996-09-10 Interim Design Inc. Interactive audiovisual distribution system
US6028626A (en) * 1995-01-03 2000-02-22 Arc Incorporated Abnormality detection and surveillance system
CA2211079A1 (en) * 1995-01-17 1996-07-25 David Sarnoff Research Center, Inc. Method and apparatus for detecting object movement within an image sequence
US5751345A (en) * 1995-02-10 1998-05-12 Dozier Financial Corporation Image retention and information security system
US6675384B1 (en) * 1995-12-21 2004-01-06 Robert S. Block Method and apparatus for information labeling and control
US5796439A (en) * 1995-12-21 1998-08-18 Siemens Medical Systems, Inc. Video format conversion process and apparatus
US5913161A (en) * 1996-04-09 1999-06-15 Adc Telecommunications, Inc. Apparatus and methods for the lawful intercept of cellular communications
US5742349A (en) * 1996-05-07 1998-04-21 Chrontel, Inc. Memory efficient video graphics subsystem with vertical filtering and scan rate conversion
US6081606A (en) * 1996-06-17 2000-06-27 Sarnoff Corporation Apparatus and a method for detecting motion within an image sequence
US5828402A (en) * 1996-06-19 1998-10-27 Canadian V-Chip Design Inc. Method and apparatus for selectively blocking audio and video signals
US7304662B1 (en) 1996-07-10 2007-12-04 Visilinx Inc. Video surveillance system and method
US5895453A (en) * 1996-08-27 1999-04-20 Sts Systems, Ltd. Method and system for the detection, management and prevention of losses in retail and other environments
US5790096A (en) * 1996-09-03 1998-08-04 Allus Technology Corporation Automated flat panel display control system for accomodating broad range of video types and formats
GB9620082D0 (en) * 1996-09-26 1996-11-13 Eyretel Ltd Signal monitoring apparatus
US20050010475A1 (en) * 1996-10-25 2005-01-13 Ipf, Inc. Internet-based brand management and marketing communication instrumentation network for deploying, installing and remotely programming brand-building server-side driven multi-mode virtual Kiosks on the World Wide Web (WWW), and methods of brand marketing communication between brand marketers and consumers using the same
US6031573A (en) * 1996-10-31 2000-02-29 Sensormatic Electronics Corporation Intelligent video information management system performing multiple functions in parallel
US6037991A (en) * 1996-11-26 2000-03-14 Motorola, Inc. Method and apparatus for communicating video information in a communication system
EP0858066A1 (en) * 1997-02-03 1998-08-12 Koninklijke Philips Electronics N.V. Method and device for converting the digital image rate
US6295367B1 (en) * 1997-06-19 2001-09-25 Emtera Corporation System and method for tracking movement of objects in a scene using correspondence graphs
US6092197A (en) * 1997-12-31 2000-07-18 The Customer Logic Company, Llc System and method for the secure discovery, exploitation and publication of information
US6014647A (en) * 1997-07-08 2000-01-11 Nizzari; Marcia M. Customer interaction tracking
US6097429A (en) * 1997-08-01 2000-08-01 Esco Electronics Corporation Site control unit for video security system
US6108711A (en) * 1998-09-11 2000-08-22 Genesys Telecommunications Laboratories, Inc. Operating system having external media layer, workflow layer, internal media layer, and knowledge base for routing media events between transactions
US6166780A (en) * 1997-10-21 2000-12-26 Principle Solutions, Inc. Automated language filter
US7139031B1 (en) * 1997-10-21 2006-11-21 Principle Solutions, Inc. Automated language filter for TV receiver
US6111610A (en) * 1997-12-11 2000-08-29 Faroudja Laboratories, Inc. Displaying film-originated video on high frame rate monitors without motions discontinuities
US6704409B1 (en) * 1997-12-31 2004-03-09 Aspect Communications Corporation Method and apparatus for processing real-time transactions and non-real-time transactions
US6327343B1 (en) * 1998-01-16 2001-12-04 International Business Machines Corporation System and methods for automatic call and data transfer processing
US6167395A (en) * 1998-09-11 2000-12-26 Genesys Telecommunications Laboratories, Inc Method and apparatus for creating specialized multimedia threads in a multimedia communication center
US6138139A (en) * 1998-10-29 2000-10-24 Genesys Telecommunications Laboraties, Inc. Method and apparatus for supporting diverse interaction paths within a multimedia communication center
US6212178B1 (en) * 1998-09-11 2001-04-03 Genesys Telecommunication Laboratories, Inc. Method and apparatus for selectively presenting media-options to clients of a multimedia call center
US6170011B1 (en) * 1998-09-11 2001-01-02 Genesys Telecommunications Laboratories, Inc. Method and apparatus for determining and initiating interaction directionality within a multimedia communication center
US6134530A (en) * 1998-04-17 2000-10-17 Andersen Consulting Llp Rule based routing system and method for a virtual sales and service center
US6070142A (en) * 1998-04-17 2000-05-30 Andersen Consulting Llp Virtual customer sales and service center and method
US20010043697A1 (en) * 1998-05-11 2001-11-22 Patrick M. Cox Monitoring of and remote access to call center activity
US6604108B1 (en) * 1998-06-05 2003-08-05 Metasolutions, Inc. Information mart system and information mart browser
US20030025599A1 (en) * 2001-05-11 2003-02-06 Monroe David A. Method and apparatus for collecting, sending, archiving and retrieving motion video and still images and notification of detected events
US6628835B1 (en) * 1998-08-31 2003-09-30 Texas Instruments Incorporated Method and system for defining and recognizing complex events in a video sequence
US6570608B1 (en) * 1998-09-30 2003-05-27 Texas Instruments Incorporated System and method for detecting interactions of people and vehicles
US6549613B1 (en) * 1998-11-05 2003-04-15 Ulysses Holding Llc Method and apparatus for intercept of wireline communications
US6330025B1 (en) * 1999-05-10 2001-12-11 Nice Systems Ltd. Digital video logging system
US7133537B1 (en) 1999-05-28 2006-11-07 It Brokerage Services Pty Limited Method and apparatus for tracking a moving object
US7103806B1 (en) * 1999-06-04 2006-09-05 Microsoft Corporation System for performing context-sensitive decisions about ideal communication modalities considering information about channel reliability
US6246752B1 (en) * 1999-06-08 2001-06-12 Valerie Bscheider System and method for data recording
GB2352948B (en) 1999-07-13 2004-03-31 Racal Recorders Ltd Voice activity monitoring apparatus and methods
US7222075B2 (en) * 1999-08-31 2007-05-22 Accenture Llp Detecting emotions using voice signal analysis
US6427137B2 (en) * 1999-08-31 2002-07-30 Accenture Llp System, method and article of manufacture for a voice analysis system that detects nervousness for preventing fraud
US6275806B1 (en) * 1999-08-31 2001-08-14 Andersen Consulting, Llp System method and article of manufacture for detecting emotion in voice signals by utilizing statistics for voice signal parameters
WO2001047222A2 (en) * 1999-12-23 2001-06-28 Ericsson Inc. Transparent communication interception in a core transport network
JP3073853U (en) * 2000-01-07 2000-12-15 船井電機株式会社 Audio mute device
US20020009000A1 (en) * 2000-01-18 2002-01-24 Qdesign Usa, Inc. Adding imperceptible noise to audio and other types of signals to cause significant degradation when compressed and decompressed
US6724887B1 (en) * 2000-01-24 2004-04-20 Verint Systems, Inc. Method and system for analyzing customer communications with a contact center
US20010052081A1 (en) * 2000-04-07 2001-12-13 Mckibben Bernard R. Communication network with a service agent element and method for providing surveillance services
US6587596B1 (en) * 2000-04-28 2003-07-01 Shutterfly, Inc. System and method of cropping an image
JP2001326632A (en) * 2000-05-17 2001-11-22 Fujitsu Ltd Distribution group management system and method
JP2001357484A (en) * 2000-06-14 2001-12-26 Kddi Corp Road abnormality detector
US6981000B2 (en) * 2000-06-30 2005-12-27 Lg Electronics Inc. Customer relationship management system and operation method thereof
AU2001284368B2 (en) 2000-08-24 2007-05-17 Nice Systems Ltd. System and method for capturing browser sessions and user actions
US20020059283A1 (en) * 2000-10-20 2002-05-16 Enteractllc Method and system for managing customer relations
WO2002037856A1 (en) 2000-11-06 2002-05-10 Dynapel Systems, Inc. Surveillance video camera enhancement system
US6441734B1 (en) 2000-12-12 2002-08-27 Koninklijke Philips Electronics N.V. Intruder detection through trajectory analysis in monitoring and surveillance systems
US20020087385A1 (en) * 2000-12-28 2002-07-04 Vincent Perry G. System and method for suggesting interaction strategies to a customer service representative
US7346186B2 (en) * 2001-01-30 2008-03-18 Nice Systems Ltd Video and audio content analysis system
WO2002093838A1 (en) * 2001-05-16 2002-11-21 Nokia Corporation Method and system allowing lawful interception of connections such a voice-over-internet-protocol calls
US7953219B2 (en) * 2001-07-19 2011-05-31 Nice Systems, Ltd. Method apparatus and system for capturing and analyzing interaction based content
GB0118921D0 (en) 2001-08-02 2001-09-26 Eyretel Telecommunications interaction analysis
US20030050834A1 (en) * 2001-09-07 2003-03-13 Sergio Caplan System and method for dynamic customizable interactive portal active during select computer time
US6912272B2 (en) * 2001-09-21 2005-06-28 Talkflow Systems, Llc Method and apparatus for managing communications and for creating communication routing rules
US20030128099A1 (en) * 2001-09-26 2003-07-10 Cockerham John M. System and method for securing a defined perimeter using multi-layered biometric electronic processing
US6559769B2 (en) * 2001-10-01 2003-05-06 Eric Anthony Early warning real-time security system
CN100592731C (en) * 2001-12-07 2010-02-24 艾利森电话股份有限公司 Lawful interception of end-to-end encrypted data traffic
WO2003067884A1 (en) 2002-02-06 2003-08-14 Nice Systems Ltd. Method and apparatus for video frame sequence-based object tracking
EP1472869A4 (en) 2002-02-06 2008-07-30 Nice Systems Ltd System and method for video content analysis-based detection, surveillance and alarm management
US7386113B2 (en) * 2002-02-25 2008-06-10 Genesys Telecommunications Laboratories, Inc. System and method for integrated resource scheduling and agent work management
US7185017B1 (en) * 2002-04-10 2007-02-27 Compuware Corporation System and method for selectively processing data sub-segments using a data mask
US7200757B1 (en) * 2002-05-13 2007-04-03 University Of Kentucky Research Foundation Data shuffling procedure for masking data
AU2003257054A1 (en) * 2002-08-16 2004-03-03 Nuasis Corporation Escalated handling of non-realtime communications
US7143288B2 (en) * 2002-10-16 2006-11-28 Vormetric, Inc. Secure file system server architecture and methods
US7076427B2 (en) * 2002-10-18 2006-07-11 Ser Solutions, Inc. Methods and apparatus for audio data monitoring and evaluation using speech recognition
US20040098295A1 (en) * 2002-11-15 2004-05-20 Iex Corporation Method and system for scheduling workload
EP1611768B1 (en) 2003-04-09 2012-02-22 TELEFONAKTIEBOLAGET LM ERICSSON (publ) Lawful interception of multimedia calls
US7447909B2 (en) 2003-06-05 2008-11-04 Nortel Networks Limited Method and system for lawful interception of packet switched network services
US7437409B2 (en) * 2003-06-13 2008-10-14 Microsoft Corporation Limiting interaction between parties in a networked session
WO2005046195A1 (en) * 2003-11-05 2005-05-19 Nice Systems Ltd. Apparatus and method for event-driven content analysis
DE10358333A1 (en) 2003-12-12 2005-07-14 Siemens Ag Telecommunication monitoring procedure uses speech and voice characteristic recognition to select communications from target user groups
WO2006021943A1 (en) * 2004-08-09 2006-03-02 Nice Systems Ltd. Apparatus and method for multimedia content based
WO2006045102A2 (en) * 2004-10-20 2006-04-27 Seven Networks, Inc. Method and apparatus for intercepting events in a communication system
US20060287617A1 (en) * 2005-06-20 2006-12-21 Department Of Veterans Affairs Autocite workstation and systems and methods therefor

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050033760A1 (en) * 1998-09-01 2005-02-10 Charles Fuller Embedded metadata engines in digital capture devices
US6843110B2 (en) * 2002-06-25 2005-01-18 Fluid Components International Llc Method and apparatus for validating the accuracy of a flowmeter
US20040267952A1 (en) * 2003-06-24 2004-12-30 He Li-Wei Variable play speed control for media streams

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7587454B2 (en) 2004-12-08 2009-09-08 Nice Systems Ltd. Video streaming parameter optimization and QoS
WO2011045116A1 (en) * 2009-10-15 2011-04-21 Thomson Licensing Method for adding voice content to video content and device for implementing said method
FR2951605A1 (en) * 2009-10-15 2011-04-22 Thomson Licensing METHOD FOR ADDING SOUND CONTENT TO VIDEO CONTENT AND DEVICE USING THE METHOD
CN102577413A (en) * 2009-10-15 2012-07-11 汤姆森特许公司 Method for adding voice content to video content and device for implementing said method
JP2013507865A (en) * 2009-10-15 2013-03-04 トムソン ライセンシング Method for adding audio content to video content and apparatus for implementing the method
CN102577413B (en) * 2009-10-15 2015-02-18 汤姆森特许公司 Method for adding voice content to video content and device for implementing said method
WO2016012484A1 (en) * 2014-07-23 2016-01-28 Koninklijke Philips N.V. Device and method for de-identification of information to be displayed
WO2016042190A1 (en) * 2014-09-15 2016-03-24 Tecnología Sostenible Y Responsable Sl Method and portable electronic device having a double screen
CN105939334A (en) * 2015-03-04 2016-09-14 费希尔-罗斯蒙特系统公司 Anomaly detection in industrial communications networks
CN105939334B (en) * 2015-03-04 2021-03-09 费希尔-罗斯蒙特系统公司 Anomaly detection in industrial communication networks
WO2020089917A1 (en) * 2018-11-02 2020-05-07 BriefCam Ltd. Method and system for automatic object-aware video or audio redaction
US11527265B2 (en) 2018-11-02 2022-12-13 BriefCam Ltd. Method and system for automatic object-aware video or audio redaction

Also Published As

Publication number Publication date
US7714878B2 (en) 2010-05-11
US20060028488A1 (en) 2006-02-09

Similar Documents

Publication Publication Date Title
US7714878B2 (en) Apparatus and method for multimedia content based manipulation
US9953147B2 (en) Computer-implemented system and method for correlating activity within a user interface with special information
Keval et al. “Not the usual suspects”: A study of factors reducing the effectiveness of CCTV
US9225701B2 (en) Secure communication systems and methods
Walklate et al. Changes and continuities in police responses to domestic abuse in England and Wales during the Covid-19 ‘lockdown’
CA2620046A1 (en) Recorded customer interactions and training system, method and computer program product
US9841865B2 (en) In-vehicle user interfaces for law enforcement
CN108021906A (en) A kind of binary channels and single-channel safety video image information analysis process system
US20060253531A1 (en) Communicating multimedia information to respondent endpoints
US11341749B2 (en) System and method to identify visitors and provide contextual services
CA2845843A1 (en) Secure communication systems and methods
Carter Confirm not command: examining fraudsters’ use of language to compel victim compliance in their own exploitation
Arikuma et al. Intelligent multimedia surveillance system for safer cities
Keval Effective design, configuration, and use of digital CCTV
Kuehn The Post-Snowden Era: Mass Surveillance and Privacy in New Zealand
Manning Technology, law and policing
US20220191430A1 (en) Systems and methods for application of context-based policies to video communication content
Gulzar et al. Surveillance privacy protection
Jibril Mediating electronic dangerscapes: A multimodal analysis of a state-sponsored newspaper warning advertisement in Nigeria
Haggerty Connecting the Pieces: China’s Bricolage Surveillance System.
Idris et al. Towards a Digital Policing in Developing Nations: The Nigerian Context
Griffin Surveillance, Technology, and American Conceptions of Freedom
Ismaila et al. Towards a Digital Policing in Developing Nations: The Nigerian Context.
US7577989B1 (en) Enterprise responder for emergencies and disaster
JP2021136016A (en) Method and system for identifying unauthorized image capture of secure monitor in work environment

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 10507582

Country of ref document: US

WWP Wipo information: published in national office

Ref document number: 10507582

Country of ref document: US

AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase